[ 470.819487] env[61987]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61987) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 470.819852] env[61987]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61987) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 470.819901] env[61987]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61987) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 470.820274] env[61987]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 470.912902] env[61987]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61987) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 470.922833] env[61987]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=61987) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 471.524065] env[61987]: INFO nova.virt.driver [None req-86246325-6373-411b-bb93-a1ac68fcc6e4 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 471.594899] env[61987]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 471.595288] env[61987]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 471.595498] env[61987]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61987) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 474.692511] env[61987]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-9050c031-78c4-405b-ae89-c1e28a28e553 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.708343] env[61987]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61987) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 474.708553] env[61987]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-d3db9f5e-07ef-4708-8f3e-6e64d75acad8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.748766] env[61987]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 33797. [ 474.748973] env[61987]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.154s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 474.749699] env[61987]: INFO nova.virt.vmwareapi.driver [None req-86246325-6373-411b-bb93-a1ac68fcc6e4 None None] VMware vCenter version: 7.0.3 [ 474.753925] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6504697d-49dc-4923-b743-2ffe153227ec {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.771624] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff684a9-14d7-440c-8f02-5fd5ebe78243 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.777625] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be73042-d2ae-41d7-a52b-13cff40c7328 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.784308] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43e7c6bd-04af-4ed4-844e-4b93b83f51da {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.797199] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ef92fc0-b854-433e-94c4-a4d52d9e662d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.802979] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03420d55-acc8-4adf-9475-a0a00ee7878d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.833023] env[61987]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-3c32027d-c28d-4b6c-a87a-c797272cbd10 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.837908] env[61987]: DEBUG nova.virt.vmwareapi.driver [None req-86246325-6373-411b-bb93-a1ac68fcc6e4 None None] Extension org.openstack.compute already exists. {{(pid=61987) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:227}} [ 474.840644] env[61987]: INFO nova.compute.provider_config [None req-86246325-6373-411b-bb93-a1ac68fcc6e4 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 475.344210] env[61987]: DEBUG nova.context [None req-86246325-6373-411b-bb93-a1ac68fcc6e4 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),a354cf11-ed17-44b8-a5b1-a3f5f94e83cd(cell1) {{(pid=61987) load_cells /opt/stack/nova/nova/context.py:464}} [ 475.346341] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 475.346575] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 475.347373] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 475.347804] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] Acquiring lock "a354cf11-ed17-44b8-a5b1-a3f5f94e83cd" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 475.348041] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] Lock "a354cf11-ed17-44b8-a5b1-a3f5f94e83cd" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 475.349057] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] Lock "a354cf11-ed17-44b8-a5b1-a3f5f94e83cd" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 475.370983] env[61987]: INFO dbcounter [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] Registered counter for database nova_cell0 [ 475.379056] env[61987]: INFO dbcounter [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] Registered counter for database nova_cell1 [ 475.382200] env[61987]: DEBUG oslo_db.sqlalchemy.engines [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61987) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 475.382573] env[61987]: DEBUG oslo_db.sqlalchemy.engines [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61987) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 475.387452] env[61987]: ERROR nova.db.main.api [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 475.387452] env[61987]: result = function(*args, **kwargs) [ 475.387452] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 475.387452] env[61987]: return func(*args, **kwargs) [ 475.387452] env[61987]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 475.387452] env[61987]: result = fn(*args, **kwargs) [ 475.387452] env[61987]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 475.387452] env[61987]: return f(*args, **kwargs) [ 475.387452] env[61987]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 475.387452] env[61987]: return db.service_get_minimum_version(context, binaries) [ 475.387452] env[61987]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 475.387452] env[61987]: _check_db_access() [ 475.387452] env[61987]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 475.387452] env[61987]: stacktrace = ''.join(traceback.format_stack()) [ 475.387452] env[61987]: [ 475.388249] env[61987]: ERROR nova.db.main.api [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 475.388249] env[61987]: result = function(*args, **kwargs) [ 475.388249] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 475.388249] env[61987]: return func(*args, **kwargs) [ 475.388249] env[61987]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 475.388249] env[61987]: result = fn(*args, **kwargs) [ 475.388249] env[61987]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 475.388249] env[61987]: return f(*args, **kwargs) [ 475.388249] env[61987]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 475.388249] env[61987]: return db.service_get_minimum_version(context, binaries) [ 475.388249] env[61987]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 475.388249] env[61987]: _check_db_access() [ 475.388249] env[61987]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 475.388249] env[61987]: stacktrace = ''.join(traceback.format_stack()) [ 475.388249] env[61987]: [ 475.388645] env[61987]: WARNING nova.objects.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 475.388800] env[61987]: WARNING nova.objects.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] Failed to get minimum service version for cell a354cf11-ed17-44b8-a5b1-a3f5f94e83cd [ 475.389232] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] Acquiring lock "singleton_lock" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 475.389397] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] Acquired lock "singleton_lock" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 475.389641] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] Releasing lock "singleton_lock" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 475.389959] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] Full set of CONF: {{(pid=61987) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 475.390153] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ******************************************************************************** {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 475.390290] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] Configuration options gathered from: {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 475.390429] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 475.390622] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 475.390750] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ================================================================================ {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 475.390957] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] allow_resize_to_same_host = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.391139] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] arq_binding_timeout = 300 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.391271] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] backdoor_port = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.391397] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] backdoor_socket = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.391561] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] block_device_allocate_retries = 60 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.391721] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] block_device_allocate_retries_interval = 3 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.391886] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cert = self.pem {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.392069] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.392271] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] compute_monitors = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.392426] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] config_dir = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.392609] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] config_drive_format = iso9660 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.392744] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.392907] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] config_source = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.393085] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] console_host = devstack {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.393256] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] control_exchange = nova {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.393444] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cpu_allocation_ratio = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.393610] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] daemon = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.393782] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] debug = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.393942] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] default_access_ip_network_name = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.394124] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] default_availability_zone = nova {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.394283] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] default_ephemeral_format = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.394442] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] default_green_pool_size = 1000 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.394678] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.394843] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] default_schedule_zone = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.395007] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] disk_allocation_ratio = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.395172] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] enable_new_services = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.395350] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] enabled_apis = ['osapi_compute'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.395515] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] enabled_ssl_apis = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.395673] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] flat_injected = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.395828] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] force_config_drive = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.395980] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] force_raw_images = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.396159] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] graceful_shutdown_timeout = 5 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.396321] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] heal_instance_info_cache_interval = 60 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.396531] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] host = cpu-1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.396711] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.396875] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] initial_disk_allocation_ratio = 1.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.397045] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] initial_ram_allocation_ratio = 1.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.397263] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.397429] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] instance_build_timeout = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.397590] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] instance_delete_interval = 300 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.397756] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] instance_format = [instance: %(uuid)s] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.397926] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] instance_name_template = instance-%08x {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.398101] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] instance_usage_audit = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.398277] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] instance_usage_audit_period = month {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.398445] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.398611] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] instances_path = /opt/stack/data/nova/instances {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.398777] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] internal_service_availability_zone = internal {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.398935] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] key = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.399105] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] live_migration_retry_count = 30 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.399275] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] log_color = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.399441] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] log_config_append = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.399610] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.399769] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] log_dir = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.399926] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] log_file = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.400061] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] log_options = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.400227] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] log_rotate_interval = 1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.400392] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] log_rotate_interval_type = days {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.400561] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] log_rotation_type = none {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.400691] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.400819] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.400985] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.401165] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.401295] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.401458] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] long_rpc_timeout = 1800 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.401617] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] max_concurrent_builds = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.401776] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] max_concurrent_live_migrations = 1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.401934] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] max_concurrent_snapshots = 5 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.402102] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] max_local_block_devices = 3 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.402265] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] max_logfile_count = 30 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.402443] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] max_logfile_size_mb = 200 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.402609] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] maximum_instance_delete_attempts = 5 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.402782] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] metadata_listen = 0.0.0.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.402951] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] metadata_listen_port = 8775 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.403135] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] metadata_workers = 2 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.403297] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] migrate_max_retries = -1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.403489] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] mkisofs_cmd = genisoimage {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.403706] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] my_block_storage_ip = 10.180.1.21 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.403841] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] my_ip = 10.180.1.21 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.404062] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] my_shared_fs_storage_ip = 10.180.1.21 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.404233] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] network_allocate_retries = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.404420] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.404592] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] osapi_compute_listen = 0.0.0.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.404763] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] osapi_compute_listen_port = 8774 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.404929] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] osapi_compute_unique_server_name_scope = {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.405113] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] osapi_compute_workers = 2 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.405280] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] password_length = 12 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.405443] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] periodic_enable = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.405609] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] periodic_fuzzy_delay = 60 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.405780] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] pointer_model = usbtablet {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.405951] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] preallocate_images = none {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.406130] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] publish_errors = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.406265] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] pybasedir = /opt/stack/nova {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.406426] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ram_allocation_ratio = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.406590] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] rate_limit_burst = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.406761] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] rate_limit_except_level = CRITICAL {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.406922] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] rate_limit_interval = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.407095] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] reboot_timeout = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.407260] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] reclaim_instance_interval = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.407420] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] record = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.407591] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] reimage_timeout_per_gb = 60 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.407760] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] report_interval = 120 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.407922] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] rescue_timeout = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.408093] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] reserved_host_cpus = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.408259] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] reserved_host_disk_mb = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.408419] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] reserved_host_memory_mb = 512 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.408583] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] reserved_huge_pages = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.408741] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] resize_confirm_window = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.408901] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] resize_fs_using_block_device = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.409069] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] resume_guests_state_on_host_boot = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.409242] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.409427] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] rpc_response_timeout = 60 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.409564] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] run_external_periodic_tasks = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.409735] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] running_deleted_instance_action = reap {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.409896] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] running_deleted_instance_poll_interval = 1800 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.410068] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] running_deleted_instance_timeout = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.410226] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] scheduler_instance_sync_interval = 120 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.410396] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] service_down_time = 720 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.410571] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] servicegroup_driver = db {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.410730] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] shell_completion = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.410892] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] shelved_offload_time = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.411066] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] shelved_poll_interval = 3600 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.411240] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] shutdown_timeout = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.411405] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] source_is_ipv6 = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.411567] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ssl_only = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.411873] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.412064] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] sync_power_state_interval = 600 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.412234] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] sync_power_state_pool_size = 1000 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.412434] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] syslog_log_facility = LOG_USER {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.412608] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] tempdir = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.412775] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] timeout_nbd = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.412946] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] transport_url = **** {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.413123] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] update_resources_interval = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.413290] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] use_cow_images = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.413476] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] use_eventlog = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.413645] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] use_journal = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.413805] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] use_json = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.413974] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] use_rootwrap_daemon = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.414155] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] use_stderr = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.414317] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] use_syslog = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.414478] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vcpu_pin_set = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.414648] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vif_plugging_is_fatal = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.414817] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vif_plugging_timeout = 300 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.414985] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] virt_mkfs = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.415162] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] volume_usage_poll_interval = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.415325] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] watch_log_file = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.415520] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] web = /usr/share/spice-html5 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 475.415712] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_concurrency.disable_process_locking = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.416014] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.416205] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.416375] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.416555] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.416728] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.416893] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.417426] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.auth_strategy = keystone {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.417426] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.compute_link_prefix = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.417426] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.417592] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.dhcp_domain = novalocal {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.417759] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.enable_instance_password = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.417928] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.glance_link_prefix = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.418106] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.418284] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.418448] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.instance_list_per_project_cells = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.418612] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.list_records_by_skipping_down_cells = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.418775] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.local_metadata_per_cell = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.418945] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.max_limit = 1000 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.419190] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.metadata_cache_expiration = 15 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.419302] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.neutron_default_tenant_id = default {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.419476] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.response_validation = warn {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.419646] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.use_neutron_default_nets = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.419823] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.419989] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.420167] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.420344] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.420517] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.vendordata_dynamic_targets = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.420684] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.vendordata_jsonfile_path = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.420866] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.421069] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.backend = dogpile.cache.memcached {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.421245] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.backend_argument = **** {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.421420] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.config_prefix = cache.oslo {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.421593] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.dead_timeout = 60.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.421757] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.debug_cache_backend = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.421921] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.enable_retry_client = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.422094] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.enable_socket_keepalive = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.422271] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.enabled = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.422465] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.enforce_fips_mode = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.422633] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.expiration_time = 600 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.422798] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.hashclient_retry_attempts = 2 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.422961] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.hashclient_retry_delay = 1.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.423137] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.memcache_dead_retry = 300 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.423300] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.memcache_password = **** {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.423519] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.423698] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.423864] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.memcache_pool_maxsize = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.424039] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.424210] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.memcache_sasl_enabled = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.424392] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.424562] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.memcache_socket_timeout = 1.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.424722] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.memcache_username = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.424889] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.proxies = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.425065] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.redis_db = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.425233] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.redis_password = **** {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.425407] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.redis_sentinel_service_name = mymaster {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.425584] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.425754] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.redis_server = localhost:6379 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.425919] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.redis_socket_timeout = 1.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.426090] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.redis_username = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.426258] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.retry_attempts = 2 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.426425] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.retry_delay = 0.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.426591] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.socket_keepalive_count = 1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.426756] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.socket_keepalive_idle = 1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.426920] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.socket_keepalive_interval = 1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.427094] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.tls_allowed_ciphers = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.427297] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.tls_cafile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.427439] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.tls_certfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.427622] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.tls_enabled = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.427785] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cache.tls_keyfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.427958] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cinder.auth_section = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.428150] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cinder.auth_type = password {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.428318] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cinder.cafile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.428496] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cinder.catalog_info = volumev3::publicURL {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.428662] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cinder.certfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.428824] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cinder.collect_timing = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.428987] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cinder.cross_az_attach = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.429170] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cinder.debug = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.429329] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cinder.endpoint_template = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.429494] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cinder.http_retries = 3 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.429657] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cinder.insecure = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.429843] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cinder.keyfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.429988] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cinder.os_region_name = RegionOne {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.430169] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cinder.split_loggers = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.430352] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cinder.timeout = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.430513] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.430675] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] compute.cpu_dedicated_set = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.430836] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] compute.cpu_shared_set = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.431019] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] compute.image_type_exclude_list = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.431182] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.431347] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] compute.max_concurrent_disk_ops = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.431512] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] compute.max_disk_devices_to_attach = -1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.431675] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.431845] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.432014] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] compute.resource_provider_association_refresh = 300 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.432181] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.432376] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] compute.shutdown_retry_interval = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.432558] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.432744] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] conductor.workers = 2 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.432925] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] console.allowed_origins = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.433102] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] console.ssl_ciphers = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.433277] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] console.ssl_minimum_version = default {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.433475] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] consoleauth.enforce_session_timeout = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.433655] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] consoleauth.token_ttl = 600 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.433828] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cyborg.cafile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.433990] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cyborg.certfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.434171] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cyborg.collect_timing = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.434332] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cyborg.connect_retries = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.434492] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cyborg.connect_retry_delay = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.434663] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cyborg.endpoint_override = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.434846] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cyborg.insecure = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.435025] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cyborg.keyfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.435186] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cyborg.max_version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.435346] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cyborg.min_version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.435509] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cyborg.region_name = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.435666] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cyborg.retriable_status_codes = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.435824] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cyborg.service_name = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.435993] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cyborg.service_type = accelerator {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.436171] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cyborg.split_loggers = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.436331] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cyborg.status_code_retries = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.436491] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cyborg.status_code_retry_delay = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.436651] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cyborg.timeout = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.436832] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.436994] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] cyborg.version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.437188] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] database.backend = sqlalchemy {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.437412] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] database.connection = **** {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.437526] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] database.connection_debug = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.437697] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] database.connection_parameters = {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.437863] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] database.connection_recycle_time = 3600 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.438034] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] database.connection_trace = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.438199] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] database.db_inc_retry_interval = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.438363] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] database.db_max_retries = 20 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.438526] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] database.db_max_retry_interval = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.438688] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] database.db_retry_interval = 1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.438847] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] database.max_overflow = 50 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.439021] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] database.max_pool_size = 5 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.439187] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] database.max_retries = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.439356] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.439542] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] database.mysql_wsrep_sync_wait = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.439710] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] database.pool_timeout = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.439875] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] database.retry_interval = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.440047] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] database.slave_connection = **** {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.440214] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] database.sqlite_synchronous = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.440414] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] database.use_db_reconnect = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.440553] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api_database.backend = sqlalchemy {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.440727] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api_database.connection = **** {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.440891] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api_database.connection_debug = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.441071] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api_database.connection_parameters = {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.441239] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api_database.connection_recycle_time = 3600 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.441400] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api_database.connection_trace = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.441563] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api_database.db_inc_retry_interval = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.441725] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api_database.db_max_retries = 20 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.441884] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api_database.db_max_retry_interval = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.442054] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api_database.db_retry_interval = 1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.442222] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api_database.max_overflow = 50 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.442414] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api_database.max_pool_size = 5 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.442581] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api_database.max_retries = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.442750] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.442910] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.443079] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api_database.pool_timeout = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.443245] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api_database.retry_interval = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.443428] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api_database.slave_connection = **** {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.443603] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] api_database.sqlite_synchronous = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.443782] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] devices.enabled_mdev_types = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.443961] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.444155] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ephemeral_storage_encryption.default_format = luks {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.444322] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ephemeral_storage_encryption.enabled = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.444487] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.444661] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.api_servers = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.444828] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.cafile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.444986] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.certfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.445163] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.collect_timing = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.445333] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.connect_retries = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.445480] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.connect_retry_delay = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.445642] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.debug = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.445808] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.default_trusted_certificate_ids = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.445969] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.enable_certificate_validation = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.446140] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.enable_rbd_download = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.446300] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.endpoint_override = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.446498] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.insecure = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.446632] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.keyfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.446791] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.max_version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.446947] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.min_version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.447124] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.num_retries = 3 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.447294] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.rbd_ceph_conf = {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.447457] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.rbd_connect_timeout = 5 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.447632] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.rbd_pool = {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.447799] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.rbd_user = {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.447959] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.region_name = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.448131] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.retriable_status_codes = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.448292] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.service_name = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.448462] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.service_type = image {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.448626] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.split_loggers = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.448786] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.status_code_retries = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.448942] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.status_code_retry_delay = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.449121] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.timeout = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.449303] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.449469] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.verify_glance_signatures = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.449631] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] glance.version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.449798] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] guestfs.debug = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.449964] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.auth_section = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.450140] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.auth_type = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.450301] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.cafile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.450465] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.certfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.450626] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.collect_timing = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.450783] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.connect_retries = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.450941] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.connect_retry_delay = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.451111] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.endpoint_override = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.451273] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.insecure = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.451431] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.keyfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.451616] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.max_version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.451779] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.min_version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.451939] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.region_name = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.452112] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.retriable_status_codes = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.452272] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.service_name = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.452463] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.service_type = shared-file-system {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.452633] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.share_apply_policy_timeout = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.452796] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.split_loggers = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.452954] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.status_code_retries = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.453123] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.status_code_retry_delay = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.453283] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.timeout = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.453462] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.453624] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] manila.version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.453792] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] mks.enabled = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.454162] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.454356] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] image_cache.manager_interval = 2400 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.454529] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] image_cache.precache_concurrency = 1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.454700] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] image_cache.remove_unused_base_images = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.454866] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.455040] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.455223] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] image_cache.subdirectory_name = _base {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.455402] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.api_max_retries = 60 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.455571] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.api_retry_interval = 2 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.455731] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.auth_section = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.455891] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.auth_type = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.456065] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.cafile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.456229] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.certfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.456394] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.collect_timing = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.456561] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.conductor_group = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.456719] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.connect_retries = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.456878] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.connect_retry_delay = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.457048] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.endpoint_override = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.457215] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.insecure = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.457376] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.keyfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.457536] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.max_version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.457692] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.min_version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.457857] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.peer_list = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.458023] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.region_name = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.458187] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.retriable_status_codes = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.458350] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.serial_console_state_timeout = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.458511] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.service_name = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.458680] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.service_type = baremetal {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.458841] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.shard = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.459014] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.split_loggers = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.459178] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.status_code_retries = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.459337] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.status_code_retry_delay = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.459498] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.timeout = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.459678] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.459839] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ironic.version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.460032] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.460213] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] key_manager.fixed_key = **** {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.460398] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.460564] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican.barbican_api_version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.460726] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican.barbican_endpoint = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.460899] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican.barbican_endpoint_type = public {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.461070] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican.barbican_region_name = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.461234] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican.cafile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.461392] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican.certfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.461554] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican.collect_timing = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.461715] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican.insecure = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.461871] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican.keyfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.462044] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican.number_of_retries = 60 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.462209] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican.retry_delay = 1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.462389] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican.send_service_user_token = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.462557] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican.split_loggers = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.462715] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican.timeout = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.462874] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican.verify_ssl = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.463042] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican.verify_ssl_path = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.463210] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican_service_user.auth_section = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.463410] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican_service_user.auth_type = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.463584] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican_service_user.cafile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.463747] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican_service_user.certfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.463911] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican_service_user.collect_timing = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.464085] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican_service_user.insecure = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.464249] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican_service_user.keyfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.464413] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican_service_user.split_loggers = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.464572] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] barbican_service_user.timeout = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.464741] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vault.approle_role_id = **** {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.464898] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vault.approle_secret_id = **** {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.465081] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vault.kv_mountpoint = secret {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.465246] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vault.kv_path = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.465418] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vault.kv_version = 2 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.465575] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vault.namespace = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.465733] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vault.root_token_id = **** {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.465891] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vault.ssl_ca_crt_file = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.466069] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vault.timeout = 60.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.466235] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vault.use_ssl = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.466405] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.466581] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.auth_section = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.466744] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.auth_type = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.466904] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.cafile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.467073] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.certfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.467238] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.collect_timing = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.467398] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.connect_retries = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.467554] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.connect_retry_delay = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.467722] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.endpoint_override = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.467863] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.insecure = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.468027] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.keyfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.468181] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.max_version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.468336] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.min_version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.468494] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.region_name = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.468651] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.retriable_status_codes = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.468809] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.service_name = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.468976] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.service_type = identity {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.469173] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.split_loggers = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.469308] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.status_code_retries = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.469469] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.status_code_retry_delay = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.469675] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.timeout = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.469894] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.470085] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] keystone.version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.470308] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.ceph_mount_options = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.470982] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.ceph_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.471196] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.connection_uri = {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.471372] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.cpu_mode = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.471547] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.cpu_model_extra_flags = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.471722] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.cpu_models = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.471896] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.cpu_power_governor_high = performance {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.472084] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.cpu_power_governor_low = powersave {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.472255] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.cpu_power_management = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.472466] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.472640] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.device_detach_attempts = 8 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.472804] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.device_detach_timeout = 20 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.472970] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.disk_cachemodes = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.473148] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.disk_prefix = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.473326] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.enabled_perf_events = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.473515] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.file_backed_memory = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.473689] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.gid_maps = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.473854] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.hw_disk_discard = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.474027] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.hw_machine_type = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.474206] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.images_rbd_ceph_conf = {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.474375] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.474551] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.474718] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.images_rbd_glance_store_name = {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.474888] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.images_rbd_pool = rbd {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.475074] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.images_type = default {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.475236] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.images_volume_group = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.475418] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.inject_key = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.475604] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.inject_partition = -2 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.475770] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.inject_password = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.475936] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.iscsi_iface = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.476112] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.iser_use_multipath = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.476281] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.live_migration_bandwidth = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.476447] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.476614] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.live_migration_downtime = 500 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.476779] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.476943] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.477116] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.live_migration_inbound_addr = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.477281] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.477444] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.live_migration_permit_post_copy = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.477603] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.live_migration_scheme = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.477774] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.live_migration_timeout_action = abort {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.477964] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.live_migration_tunnelled = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.478105] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.live_migration_uri = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.478269] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.live_migration_with_native_tls = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.478432] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.max_queues = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.478596] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.478827] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.478993] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.nfs_mount_options = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.479294] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.479472] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.479639] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.num_iser_scan_tries = 5 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.479803] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.num_memory_encrypted_guests = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.479970] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.480147] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.num_pcie_ports = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.480324] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.num_volume_scan_tries = 5 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.480489] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.pmem_namespaces = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.480650] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.quobyte_client_cfg = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.480936] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.481124] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.rbd_connect_timeout = 5 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.481292] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.481460] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.481633] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.rbd_secret_uuid = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.481781] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.rbd_user = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.481945] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.482129] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.remote_filesystem_transport = ssh {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.482294] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.rescue_image_id = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.482481] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.rescue_kernel_id = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.482651] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.rescue_ramdisk_id = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.482821] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.482982] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.rx_queue_size = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.483163] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.smbfs_mount_options = {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.483483] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.483676] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.snapshot_compression = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.483843] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.snapshot_image_format = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.484082] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.484257] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.sparse_logical_volumes = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.484423] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.swtpm_enabled = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.484595] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.swtpm_group = tss {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.484763] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.swtpm_user = tss {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.484933] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.sysinfo_serial = unique {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.485108] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.tb_cache_size = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.485269] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.tx_queue_size = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.485433] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.uid_maps = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.485598] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.use_virtio_for_bridges = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.485770] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.virt_type = kvm {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.485939] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.volume_clear = zero {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.486117] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.volume_clear_size = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.486287] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.volume_use_multipath = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.486448] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.vzstorage_cache_path = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.486619] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.486786] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.vzstorage_mount_group = qemu {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.486952] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.vzstorage_mount_opts = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.487135] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.487446] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.487889] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.vzstorage_mount_user = stack {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.487889] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.487983] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.auth_section = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.488167] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.auth_type = password {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.488331] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.cafile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.488491] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.certfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.488656] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.collect_timing = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.488812] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.connect_retries = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.488967] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.connect_retry_delay = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.489149] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.default_floating_pool = public {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.489310] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.endpoint_override = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.489473] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.extension_sync_interval = 600 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.489635] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.http_retries = 3 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.489795] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.insecure = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.489953] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.keyfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.490123] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.max_version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.490295] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.490459] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.min_version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.490628] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.ovs_bridge = br-int {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.490792] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.physnets = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.490990] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.region_name = RegionOne {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.491130] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.retriable_status_codes = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.491301] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.service_metadata_proxy = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.491460] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.service_name = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.491630] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.service_type = network {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.491792] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.split_loggers = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.491947] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.status_code_retries = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.492119] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.status_code_retry_delay = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.492278] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.timeout = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.492487] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.492660] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] neutron.version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.492834] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] notifications.bdms_in_notifications = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.493085] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] notifications.default_level = INFO {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.493280] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] notifications.include_share_mapping = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.493490] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] notifications.notification_format = unversioned {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.493668] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] notifications.notify_on_state_change = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.493849] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.494042] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] pci.alias = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.494220] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] pci.device_spec = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.494387] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] pci.report_in_placement = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.494562] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.auth_section = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.494735] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.auth_type = password {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.494907] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.495090] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.cafile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.495277] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.certfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.495419] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.collect_timing = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.495581] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.connect_retries = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.495740] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.connect_retry_delay = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.495900] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.default_domain_id = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.496070] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.default_domain_name = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.496233] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.domain_id = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.496391] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.domain_name = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.496550] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.endpoint_override = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.496711] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.insecure = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.496867] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.keyfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.497037] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.max_version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.497198] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.min_version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.497370] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.password = **** {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.497530] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.project_domain_id = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.497697] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.project_domain_name = Default {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.497863] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.project_id = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.498046] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.project_name = service {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.498223] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.region_name = RegionOne {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.498388] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.retriable_status_codes = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.498549] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.service_name = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.498719] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.service_type = placement {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.498884] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.split_loggers = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.499056] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.status_code_retries = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.499224] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.status_code_retry_delay = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.499386] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.system_scope = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.499570] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.timeout = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.499739] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.trust_id = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.499900] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.user_domain_id = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.500082] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.user_domain_name = Default {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.500249] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.user_id = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.500428] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.username = nova {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.500613] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.500778] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] placement.version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.500954] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] quota.cores = 20 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.501135] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] quota.count_usage_from_placement = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.501308] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.501486] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] quota.injected_file_content_bytes = 10240 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.501656] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] quota.injected_file_path_length = 255 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.501823] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] quota.injected_files = 5 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.501994] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] quota.instances = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.502175] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] quota.key_pairs = 100 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.502361] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] quota.metadata_items = 128 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.502535] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] quota.ram = 51200 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.502703] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] quota.recheck_quota = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.502870] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] quota.server_group_members = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.503047] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] quota.server_groups = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.503227] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.503411] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.503589] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] scheduler.image_metadata_prefilter = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.503757] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.503925] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] scheduler.max_attempts = 3 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.504103] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] scheduler.max_placement_results = 1000 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.504274] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.504437] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] scheduler.query_placement_for_image_type_support = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.504604] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.504780] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] scheduler.workers = 2 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.504953] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.505140] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.505373] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.505497] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.505665] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.505830] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.505993] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.506196] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.506367] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.host_subset_size = 1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.506533] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.506693] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.506854] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.507035] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.isolated_hosts = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.507204] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.isolated_images = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.507366] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.507527] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.507687] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.507851] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.pci_in_placement = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.508020] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.508190] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.508355] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.508533] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.508695] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.508856] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.509025] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.track_instance_changes = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.509208] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.509380] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] metrics.required = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.509546] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] metrics.weight_multiplier = 1.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.509708] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.509874] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] metrics.weight_setting = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.510210] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.510393] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] serial_console.enabled = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.510568] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] serial_console.port_range = 10000:20000 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.510742] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.510914] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.511096] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] serial_console.serialproxy_port = 6083 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.511269] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] service_user.auth_section = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.511442] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] service_user.auth_type = password {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.511636] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] service_user.cafile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.511803] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] service_user.certfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.511970] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] service_user.collect_timing = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.512146] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] service_user.insecure = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.512308] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] service_user.keyfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.512509] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] service_user.send_service_user_token = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.512680] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] service_user.split_loggers = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.512842] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] service_user.timeout = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.513022] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] spice.agent_enabled = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.513194] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] spice.enabled = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.513529] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.513730] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.513901] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] spice.html5proxy_port = 6082 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.514074] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] spice.image_compression = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.514237] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] spice.jpeg_compression = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.514395] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] spice.playback_compression = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.514559] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] spice.require_secure = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.514731] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] spice.server_listen = 127.0.0.1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.514898] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.515069] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] spice.streaming_mode = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.515230] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] spice.zlib_compression = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.515397] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] upgrade_levels.baseapi = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.515569] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] upgrade_levels.compute = auto {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.515728] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] upgrade_levels.conductor = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.515886] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] upgrade_levels.scheduler = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.516061] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vendordata_dynamic_auth.auth_section = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.516230] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vendordata_dynamic_auth.auth_type = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.516390] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vendordata_dynamic_auth.cafile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.516550] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vendordata_dynamic_auth.certfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.516713] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.516872] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vendordata_dynamic_auth.insecure = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.517041] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vendordata_dynamic_auth.keyfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.517213] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.517376] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vendordata_dynamic_auth.timeout = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.517549] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.api_retry_count = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.517714] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.ca_file = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.517882] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.cache_prefix = devstack-image-cache {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.518060] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.cluster_name = testcl1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.518230] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.connection_pool_size = 10 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.518425] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.console_delay_seconds = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.518553] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.datastore_regex = ^datastore.* {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.518761] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.518932] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.host_password = **** {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.519110] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.host_port = 443 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.519281] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.host_username = administrator@vsphere.local {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.519450] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.insecure = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.519615] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.integration_bridge = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.519780] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.maximum_objects = 100 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.519941] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.pbm_default_policy = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.520115] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.pbm_enabled = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.520278] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.pbm_wsdl_location = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.520449] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.520609] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.serial_port_proxy_uri = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.520766] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.serial_port_service_uri = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.520933] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.task_poll_interval = 0.5 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.521135] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.use_linked_clone = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.521314] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.vnc_keymap = en-us {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.521485] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.vnc_port = 5900 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.521653] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vmware.vnc_port_total = 10000 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.521841] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vnc.auth_schemes = ['none'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.522029] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vnc.enabled = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.522336] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.522549] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.522757] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vnc.novncproxy_port = 6080 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.522908] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vnc.server_listen = 127.0.0.1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.523093] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.523264] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vnc.vencrypt_ca_certs = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.523457] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vnc.vencrypt_client_cert = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.523630] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vnc.vencrypt_client_key = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.523811] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.523977] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] workarounds.disable_deep_image_inspection = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.524157] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.524320] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.524479] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.524644] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] workarounds.disable_rootwrap = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.524803] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] workarounds.enable_numa_live_migration = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.524961] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.525135] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.525298] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.525456] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] workarounds.libvirt_disable_apic = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.525616] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.525779] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.525939] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.526109] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.526272] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.526441] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.526604] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.526763] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.526919] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.527096] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.527281] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.527452] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] wsgi.client_socket_timeout = 900 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.527620] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] wsgi.default_pool_size = 1000 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.527786] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] wsgi.keep_alive = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.527954] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] wsgi.max_header_line = 16384 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.528130] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] wsgi.secure_proxy_ssl_header = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.528295] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] wsgi.ssl_ca_file = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.528458] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] wsgi.ssl_cert_file = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.528621] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] wsgi.ssl_key_file = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.528787] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] wsgi.tcp_keepidle = 600 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.528964] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.529146] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] zvm.ca_file = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.529309] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] zvm.cloud_connector_url = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.529601] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.529779] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] zvm.reachable_timeout = 300 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.529960] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_policy.enforce_new_defaults = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.530362] env[61987]: WARNING oslo_config.cfg [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 475.530551] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_policy.enforce_scope = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.530732] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_policy.policy_default_rule = default {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.530919] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.531108] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_policy.policy_file = policy.yaml {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.531291] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.531446] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.531610] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.531772] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.531937] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.532118] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.532298] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.532512] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] profiler.connection_string = messaging:// {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.532691] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] profiler.enabled = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.532864] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] profiler.es_doc_type = notification {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.533041] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] profiler.es_scroll_size = 10000 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.533216] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] profiler.es_scroll_time = 2m {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.533394] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] profiler.filter_error_trace = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.533595] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] profiler.hmac_keys = **** {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.533771] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] profiler.sentinel_service_name = mymaster {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.533942] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] profiler.socket_timeout = 0.1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.534120] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] profiler.trace_requests = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.534284] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] profiler.trace_sqlalchemy = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.534469] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] profiler_jaeger.process_tags = {} {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.534636] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] profiler_jaeger.service_name_prefix = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.534799] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] profiler_otlp.service_name_prefix = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.534966] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] remote_debug.host = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.535142] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] remote_debug.port = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.535321] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.535485] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.535691] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.535817] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.535983] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.536156] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.536320] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.536484] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.536648] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.536821] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.536984] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.537171] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.537342] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.537513] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.537686] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.537856] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.538032] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.538215] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.538383] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.538547] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.538714] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.538879] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.539051] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.539224] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.539388] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.539553] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.539718] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.539879] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.540056] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.540228] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.ssl = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.540402] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.540576] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.540740] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.540912] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.541096] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.ssl_version = {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.541263] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.541452] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.541625] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_notifications.retry = -1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.541808] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.541984] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_messaging_notifications.transport_url = **** {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.542171] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.auth_section = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.542363] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.auth_type = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.542539] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.cafile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.542705] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.certfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.542869] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.collect_timing = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.543037] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.connect_retries = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.543203] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.connect_retry_delay = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.543376] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.endpoint_id = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.543570] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.endpoint_override = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.543738] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.insecure = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.543894] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.keyfile = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.544066] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.max_version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.544231] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.min_version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.544389] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.region_name = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.544555] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.retriable_status_codes = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.544714] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.service_name = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.544872] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.service_type = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.545041] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.split_loggers = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.545206] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.status_code_retries = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.545366] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.status_code_retry_delay = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.545527] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.timeout = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.545685] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.valid_interfaces = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.545842] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_limit.version = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.546019] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_reports.file_event_handler = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.546186] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.546358] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] oslo_reports.log_dir = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.546553] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.546718] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.546877] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.547055] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.547226] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.547388] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.547562] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.547726] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vif_plug_ovs_privileged.group = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.547887] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.548063] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.548233] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.548394] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] vif_plug_ovs_privileged.user = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.548567] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] os_vif_linux_bridge.flat_interface = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.548791] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.548922] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.549105] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.549279] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.549473] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.549659] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.549824] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.550010] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.550197] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] os_vif_ovs.isolate_vif = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.550393] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.550569] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.550740] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.550911] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] os_vif_ovs.ovsdb_interface = native {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.551086] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] os_vif_ovs.per_port_bridge = False {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.551259] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] os_brick.lock_path = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.551428] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.551596] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.551768] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] privsep_osbrick.capabilities = [21] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.551926] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] privsep_osbrick.group = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.552095] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] privsep_osbrick.helper_command = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.552265] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.552453] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.552625] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] privsep_osbrick.user = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.552801] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.552965] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] nova_sys_admin.group = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.553137] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] nova_sys_admin.helper_command = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.553305] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.553471] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.553629] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] nova_sys_admin.user = None {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 475.553759] env[61987]: DEBUG oslo_service.service [None req-b9ef0db6-f937-4da6-b179-84bcef4502b6 None None] ******************************************************************************** {{(pid=61987) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 475.554263] env[61987]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 476.057706] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Getting list of instances from cluster (obj){ [ 476.057706] env[61987]: value = "domain-c8" [ 476.057706] env[61987]: _type = "ClusterComputeResource" [ 476.057706] env[61987]: } {{(pid=61987) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 476.058893] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df1c96d-eb30-4fee-88a9-31754bc9152d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 476.068216] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Got total of 0 instances {{(pid=61987) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 476.068758] env[61987]: WARNING nova.virt.vmwareapi.driver [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 476.069247] env[61987]: INFO nova.virt.node [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Generated node identity 8db744e5-1e14-4e98-9818-695ea4b7dfd0 [ 476.069494] env[61987]: INFO nova.virt.node [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Wrote node identity 8db744e5-1e14-4e98-9818-695ea4b7dfd0 to /opt/stack/data/n-cpu-1/compute_id [ 476.572528] env[61987]: WARNING nova.compute.manager [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Compute nodes ['8db744e5-1e14-4e98-9818-695ea4b7dfd0'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 477.578097] env[61987]: INFO nova.compute.manager [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 478.583623] env[61987]: WARNING nova.compute.manager [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 478.583999] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 478.584114] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 478.584268] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 478.584424] env[61987]: DEBUG nova.compute.resource_tracker [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61987) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 478.585384] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5edebb25-9161-4966-89ab-fbb6d8081090 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 478.593863] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9651734a-d0b4-44d2-bc00-3dbf767966aa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 478.608822] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-053a40c3-a879-4fad-85e6-7dc9ec1d419c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 478.615516] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e47eaa5-4d73-4381-95c2-92139c3b8161 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 478.644540] env[61987]: DEBUG nova.compute.resource_tracker [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181564MB free_disk=180GB free_vcpus=48 pci_devices=None {{(pid=61987) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 478.644738] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 478.644959] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 479.147769] env[61987]: WARNING nova.compute.resource_tracker [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] No compute node record for cpu-1:8db744e5-1e14-4e98-9818-695ea4b7dfd0: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 8db744e5-1e14-4e98-9818-695ea4b7dfd0 could not be found. [ 479.651911] env[61987]: INFO nova.compute.resource_tracker [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 [ 481.160282] env[61987]: DEBUG nova.compute.resource_tracker [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 481.160663] env[61987]: DEBUG nova.compute.resource_tracker [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 481.313660] env[61987]: INFO nova.scheduler.client.report [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] [req-e4fc6244-e4f0-4868-9ed8-6c85dc76086c] Created resource provider record via placement API for resource provider with UUID 8db744e5-1e14-4e98-9818-695ea4b7dfd0 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 481.330053] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c71b6448-ad0d-46fb-805f-264ca6195211 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.337931] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac750f08-7ba1-4afc-80ff-5fe3b454e0f6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.367414] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8836333f-1a39-459f-a259-63baa0088eda {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.374095] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e58f3b-4bbc-4c79-9152-ea3fc5b65dd1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 481.386631] env[61987]: DEBUG nova.compute.provider_tree [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 481.923613] env[61987]: DEBUG nova.scheduler.client.report [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Updated inventory for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 481.923885] env[61987]: DEBUG nova.compute.provider_tree [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Updating resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 generation from 0 to 1 during operation: update_inventory {{(pid=61987) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 481.924118] env[61987]: DEBUG nova.compute.provider_tree [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 481.973418] env[61987]: DEBUG nova.compute.provider_tree [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Updating resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 generation from 1 to 2 during operation: update_traits {{(pid=61987) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 482.478020] env[61987]: DEBUG nova.compute.resource_tracker [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61987) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 482.478452] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.833s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 482.478452] env[61987]: DEBUG nova.service [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Creating RPC server for service compute {{(pid=61987) start /opt/stack/nova/nova/service.py:186}} [ 482.494572] env[61987]: DEBUG nova.service [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] Join ServiceGroup membership for this service compute {{(pid=61987) start /opt/stack/nova/nova/service.py:203}} [ 482.494770] env[61987]: DEBUG nova.servicegroup.drivers.db [None req-a82331a4-6ac5-4077-99ff-2d77a232f48a None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61987) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 510.500711] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 511.007891] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Getting list of instances from cluster (obj){ [ 511.007891] env[61987]: value = "domain-c8" [ 511.007891] env[61987]: _type = "ClusterComputeResource" [ 511.007891] env[61987]: } {{(pid=61987) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 511.007891] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-635ec7ea-12dd-4e59-b646-2c051058ccae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.017153] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Got total of 0 instances {{(pid=61987) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 511.017403] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 511.017724] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Getting list of instances from cluster (obj){ [ 511.017724] env[61987]: value = "domain-c8" [ 511.017724] env[61987]: _type = "ClusterComputeResource" [ 511.017724] env[61987]: } {{(pid=61987) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 511.018772] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bb1cb74-4d4f-40c9-8039-d178c298c8ed {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.027195] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Got total of 0 instances {{(pid=61987) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 517.362618] env[61987]: DEBUG oslo_concurrency.lockutils [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "31c175f9-20a5-4e38-ba90-972b847ab65e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.362880] env[61987]: DEBUG oslo_concurrency.lockutils [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "31c175f9-20a5-4e38-ba90-972b847ab65e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.866507] env[61987]: DEBUG nova.compute.manager [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 518.178043] env[61987]: DEBUG oslo_concurrency.lockutils [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Acquiring lock "25f011ba-eb0b-4295-be8a-d223584ada63" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.178043] env[61987]: DEBUG oslo_concurrency.lockutils [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Lock "25f011ba-eb0b-4295-be8a-d223584ada63" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.413265] env[61987]: DEBUG oslo_concurrency.lockutils [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.413877] env[61987]: DEBUG oslo_concurrency.lockutils [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.421141] env[61987]: INFO nova.compute.claims [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 518.680603] env[61987]: DEBUG nova.compute.manager [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 518.904707] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Acquiring lock "7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.905025] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Lock "7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.070979] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Acquiring lock "ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.071365] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Lock "ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.213632] env[61987]: DEBUG oslo_concurrency.lockutils [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.408816] env[61987]: DEBUG nova.compute.manager [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 519.572202] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101223bc-7577-40dc-a963-7ea36d4f2426 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.578350] env[61987]: DEBUG nova.compute.manager [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 519.588265] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43d85071-2924-482d-b38b-6799e7001fb7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.623621] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e1e3e29-afb1-4d9a-8ced-a6a5864e81c5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.633840] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a748633-19c8-469e-9015-ec78c826ba97 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.649561] env[61987]: DEBUG nova.compute.provider_tree [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 519.888632] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Acquiring lock "71316e2e-cda6-461c-8004-ff80fc2da90c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.888884] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Lock "71316e2e-cda6-461c-8004-ff80fc2da90c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.942977] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.102192] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.152846] env[61987]: DEBUG nova.scheduler.client.report [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 520.394231] env[61987]: DEBUG nova.compute.manager [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 520.660580] env[61987]: DEBUG oslo_concurrency.lockutils [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.245s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 520.660580] env[61987]: DEBUG nova.compute.manager [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 520.662229] env[61987]: DEBUG oslo_concurrency.lockutils [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.449s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.664605] env[61987]: INFO nova.compute.claims [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 520.928908] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.120177] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Acquiring lock "a0a004c3-6254-4f47-a080-2b9cc85a2a1d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.120177] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Lock "a0a004c3-6254-4f47-a080-2b9cc85a2a1d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.171347] env[61987]: DEBUG nova.compute.utils [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 521.174765] env[61987]: DEBUG nova.compute.manager [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 521.175087] env[61987]: DEBUG nova.network.neutron [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 521.504813] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.504813] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.624017] env[61987]: DEBUG nova.compute.manager [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 521.692386] env[61987]: DEBUG nova.policy [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4ead7955f1d475590523ac98c579956', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '315a2a3450024ed38e27cd7758f392f9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 521.694902] env[61987]: DEBUG nova.compute.manager [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 521.836981] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73d17f8a-1fda-4b4e-9819-01c75926a313 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.845522] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c49015-0b1b-4dd5-afab-d931ebfdb421 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.878117] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-742bb12b-557f-4dc8-8f98-8b7d903984f7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.890895] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6510f25a-3b62-42a2-bc4d-89578cb033f7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.905448] env[61987]: DEBUG nova.compute.provider_tree [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 522.007382] env[61987]: DEBUG nova.compute.manager [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 522.064517] env[61987]: DEBUG nova.network.neutron [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Successfully created port: 691fbcaf-62ed-49e6-b5c0-06b24549eca4 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 522.154118] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.365911] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquiring lock "510803e9-89ce-4452-b245-ee1f124ef081" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.366149] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "510803e9-89ce-4452-b245-ee1f124ef081" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.391464] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquiring lock "1a9030ce-3a35-4760-96f4-4288c1d2f9d1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.392557] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "1a9030ce-3a35-4760-96f4-4288c1d2f9d1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.409262] env[61987]: DEBUG nova.scheduler.client.report [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 522.442371] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquiring lock "c9ce2911-8fc4-4aa8-883e-487ee06670ba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.442685] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "c9ce2911-8fc4-4aa8-883e-487ee06670ba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.523286] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Acquiring lock "7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.523543] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Lock "7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.534983] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.708049] env[61987]: DEBUG nova.compute.manager [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 522.746028] env[61987]: DEBUG nova.virt.hardware [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 522.746156] env[61987]: DEBUG nova.virt.hardware [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 522.746317] env[61987]: DEBUG nova.virt.hardware [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 522.746503] env[61987]: DEBUG nova.virt.hardware [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 522.747322] env[61987]: DEBUG nova.virt.hardware [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 522.747322] env[61987]: DEBUG nova.virt.hardware [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 522.747322] env[61987]: DEBUG nova.virt.hardware [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 522.747322] env[61987]: DEBUG nova.virt.hardware [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 522.747874] env[61987]: DEBUG nova.virt.hardware [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 522.747874] env[61987]: DEBUG nova.virt.hardware [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 522.748109] env[61987]: DEBUG nova.virt.hardware [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 522.749040] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02473467-1636-47bb-aa7e-f7c23f4dbeba {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.761142] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33675607-3154-4caf-a5a2-93105cc26963 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.787589] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eabf779-b77e-4ab6-8328-9a4dccf87726 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.802175] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquiring lock "3182ca24-c989-4c29-8e47-d913d0607240" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.802399] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "3182ca24-c989-4c29-8e47-d913d0607240" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.868126] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 522.896481] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 522.918855] env[61987]: DEBUG oslo_concurrency.lockutils [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.257s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.919487] env[61987]: DEBUG nova.compute.manager [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 522.925744] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.980s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.925744] env[61987]: INFO nova.compute.claims [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 522.946515] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 523.407741] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.424249] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.431469] env[61987]: DEBUG nova.compute.utils [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 523.437909] env[61987]: DEBUG nova.compute.manager [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 523.437909] env[61987]: DEBUG nova.network.neutron [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 523.468766] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.557143] env[61987]: DEBUG nova.policy [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6e88d7b7fc8e45efaf2cce3e44f159bc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f68fe18275274adca1f750821e7e2354', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 523.936613] env[61987]: DEBUG nova.compute.manager [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 524.246373] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b27c5051-ff5b-4e19-87ba-5cc4e5f02e45 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.254710] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1387ce07-695f-464f-b604-408a7aae8cf8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.292204] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81bfaac3-16da-4344-88bb-d49c88e91acb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.300839] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03128af8-51d3-44fc-88d0-09e7fbdadf6d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.315320] env[61987]: DEBUG nova.compute.provider_tree [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 524.571107] env[61987]: DEBUG nova.network.neutron [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Successfully created port: 3f3ca432-e7a4-4a2f-9086-3983a2f72e29 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 524.818995] env[61987]: DEBUG nova.scheduler.client.report [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 524.950767] env[61987]: DEBUG nova.compute.manager [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 524.982096] env[61987]: DEBUG nova.virt.hardware [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 524.982214] env[61987]: DEBUG nova.virt.hardware [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 524.982336] env[61987]: DEBUG nova.virt.hardware [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 524.982562] env[61987]: DEBUG nova.virt.hardware [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 524.983041] env[61987]: DEBUG nova.virt.hardware [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 524.983041] env[61987]: DEBUG nova.virt.hardware [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 524.983651] env[61987]: DEBUG nova.virt.hardware [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 524.983879] env[61987]: DEBUG nova.virt.hardware [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 524.984078] env[61987]: DEBUG nova.virt.hardware [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 524.984252] env[61987]: DEBUG nova.virt.hardware [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 524.984425] env[61987]: DEBUG nova.virt.hardware [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 524.985351] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d881697d-7728-4cfd-9e1d-1ba303577a72 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.996756] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc60401-cfb2-44d8-98e5-2f10e6eab7fc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.060987] env[61987]: ERROR nova.compute.manager [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 691fbcaf-62ed-49e6-b5c0-06b24549eca4, please check neutron logs for more information. [ 525.060987] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 525.060987] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 525.060987] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 525.060987] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 525.060987] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 525.060987] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 525.060987] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 525.060987] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 525.060987] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 525.060987] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 525.060987] env[61987]: ERROR nova.compute.manager raise self.value [ 525.060987] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 525.060987] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 525.060987] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 525.060987] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 525.061515] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 525.061515] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 525.061515] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 691fbcaf-62ed-49e6-b5c0-06b24549eca4, please check neutron logs for more information. [ 525.061515] env[61987]: ERROR nova.compute.manager [ 525.063010] env[61987]: Traceback (most recent call last): [ 525.063070] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 525.063070] env[61987]: listener.cb(fileno) [ 525.063070] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 525.063070] env[61987]: result = function(*args, **kwargs) [ 525.063070] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 525.063070] env[61987]: return func(*args, **kwargs) [ 525.063070] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 525.063070] env[61987]: raise e [ 525.063070] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 525.063070] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 525.063070] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 525.063070] env[61987]: created_port_ids = self._update_ports_for_instance( [ 525.063380] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 525.063380] env[61987]: with excutils.save_and_reraise_exception(): [ 525.063380] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 525.063380] env[61987]: self.force_reraise() [ 525.063380] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 525.063380] env[61987]: raise self.value [ 525.063380] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 525.063380] env[61987]: updated_port = self._update_port( [ 525.063380] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 525.063380] env[61987]: _ensure_no_port_binding_failure(port) [ 525.063380] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 525.063380] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 525.063380] env[61987]: nova.exception.PortBindingFailed: Binding failed for port 691fbcaf-62ed-49e6-b5c0-06b24549eca4, please check neutron logs for more information. [ 525.063380] env[61987]: Removing descriptor: 15 [ 525.065481] env[61987]: ERROR nova.compute.manager [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 691fbcaf-62ed-49e6-b5c0-06b24549eca4, please check neutron logs for more information. [ 525.065481] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Traceback (most recent call last): [ 525.065481] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 525.065481] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] yield resources [ 525.065481] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 525.065481] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] self.driver.spawn(context, instance, image_meta, [ 525.065481] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 525.065481] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 525.065481] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 525.065481] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] vm_ref = self.build_virtual_machine(instance, [ 525.065481] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 525.065820] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] vif_infos = vmwarevif.get_vif_info(self._session, [ 525.065820] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 525.065820] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] for vif in network_info: [ 525.065820] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 525.065820] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] return self._sync_wrapper(fn, *args, **kwargs) [ 525.065820] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 525.065820] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] self.wait() [ 525.065820] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 525.065820] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] self[:] = self._gt.wait() [ 525.065820] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 525.065820] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] return self._exit_event.wait() [ 525.065820] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 525.065820] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] result = hub.switch() [ 525.066541] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 525.066541] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] return self.greenlet.switch() [ 525.066541] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 525.066541] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] result = function(*args, **kwargs) [ 525.066541] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 525.066541] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] return func(*args, **kwargs) [ 525.066541] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 525.066541] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] raise e [ 525.066541] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 525.066541] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] nwinfo = self.network_api.allocate_for_instance( [ 525.066541] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 525.066541] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] created_port_ids = self._update_ports_for_instance( [ 525.066541] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 525.067071] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] with excutils.save_and_reraise_exception(): [ 525.067071] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 525.067071] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] self.force_reraise() [ 525.067071] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 525.067071] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] raise self.value [ 525.067071] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 525.067071] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] updated_port = self._update_port( [ 525.067071] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 525.067071] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] _ensure_no_port_binding_failure(port) [ 525.067071] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 525.067071] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] raise exception.PortBindingFailed(port_id=port['id']) [ 525.067071] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] nova.exception.PortBindingFailed: Binding failed for port 691fbcaf-62ed-49e6-b5c0-06b24549eca4, please check neutron logs for more information. [ 525.067071] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] [ 525.067407] env[61987]: INFO nova.compute.manager [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Terminating instance [ 525.327743] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.328093] env[61987]: DEBUG nova.compute.manager [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 525.337284] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.232s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.340485] env[61987]: INFO nova.compute.claims [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 525.573501] env[61987]: DEBUG oslo_concurrency.lockutils [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "refresh_cache-31c175f9-20a5-4e38-ba90-972b847ab65e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 525.573841] env[61987]: DEBUG oslo_concurrency.lockutils [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired lock "refresh_cache-31c175f9-20a5-4e38-ba90-972b847ab65e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 525.573942] env[61987]: DEBUG nova.network.neutron [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 525.847715] env[61987]: DEBUG nova.compute.utils [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 525.853974] env[61987]: DEBUG nova.compute.manager [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 525.856719] env[61987]: DEBUG nova.network.neutron [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 526.057401] env[61987]: DEBUG nova.policy [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '87a8bf3961ce4d0c8e6afec8d26e8c04', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7d8c827ae5be4898a21417c0084d66dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 526.162112] env[61987]: DEBUG nova.network.neutron [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 526.359018] env[61987]: DEBUG nova.compute.manager [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 526.542092] env[61987]: DEBUG nova.network.neutron [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.562671] env[61987]: DEBUG nova.compute.manager [req-4fb3a036-ef55-4b8b-a69a-329afacc7bef req-cc56c294-cb18-4335-81f6-05d7c2901249 service nova] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Received event network-changed-691fbcaf-62ed-49e6-b5c0-06b24549eca4 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 526.562671] env[61987]: DEBUG nova.compute.manager [req-4fb3a036-ef55-4b8b-a69a-329afacc7bef req-cc56c294-cb18-4335-81f6-05d7c2901249 service nova] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Refreshing instance network info cache due to event network-changed-691fbcaf-62ed-49e6-b5c0-06b24549eca4. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 526.562671] env[61987]: DEBUG oslo_concurrency.lockutils [req-4fb3a036-ef55-4b8b-a69a-329afacc7bef req-cc56c294-cb18-4335-81f6-05d7c2901249 service nova] Acquiring lock "refresh_cache-31c175f9-20a5-4e38-ba90-972b847ab65e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 526.651966] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b0458d3-f85d-4bcf-af69-649520465ecb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.662884] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-688f1ce0-da87-426a-9532-74fdfadd337d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.697417] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-635e78f6-e2de-4c35-8d39-bd1e0b93d39f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.706462] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baa6b210-019e-41fa-8d20-dbf3c9de7a10 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.722290] env[61987]: DEBUG nova.compute.provider_tree [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 527.046071] env[61987]: DEBUG oslo_concurrency.lockutils [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Releasing lock "refresh_cache-31c175f9-20a5-4e38-ba90-972b847ab65e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.047925] env[61987]: DEBUG nova.compute.manager [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 527.047925] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 527.047925] env[61987]: DEBUG oslo_concurrency.lockutils [req-4fb3a036-ef55-4b8b-a69a-329afacc7bef req-cc56c294-cb18-4335-81f6-05d7c2901249 service nova] Acquired lock "refresh_cache-31c175f9-20a5-4e38-ba90-972b847ab65e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.047925] env[61987]: DEBUG nova.network.neutron [req-4fb3a036-ef55-4b8b-a69a-329afacc7bef req-cc56c294-cb18-4335-81f6-05d7c2901249 service nova] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Refreshing network info cache for port 691fbcaf-62ed-49e6-b5c0-06b24549eca4 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 527.050346] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-27de8fbf-1f7e-4103-b3ad-72ad00b53a52 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.061136] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-296acef3-1eab-4e0a-b1e6-df38e56a281e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.093019] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 31c175f9-20a5-4e38-ba90-972b847ab65e could not be found. [ 527.093255] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 527.094022] env[61987]: INFO nova.compute.manager [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Took 0.05 seconds to destroy the instance on the hypervisor. [ 527.094022] env[61987]: DEBUG oslo.service.loopingcall [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 527.094144] env[61987]: DEBUG nova.compute.manager [-] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 527.094235] env[61987]: DEBUG nova.network.neutron [-] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 527.153182] env[61987]: DEBUG nova.network.neutron [-] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 527.207738] env[61987]: DEBUG nova.network.neutron [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Successfully created port: cd2e6086-23ab-4064-a665-7e4b92b07e6f {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 527.226273] env[61987]: DEBUG nova.scheduler.client.report [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 527.372431] env[61987]: DEBUG nova.compute.manager [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 527.417346] env[61987]: DEBUG nova.virt.hardware [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 527.417681] env[61987]: DEBUG nova.virt.hardware [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 527.418413] env[61987]: DEBUG nova.virt.hardware [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 527.418862] env[61987]: DEBUG nova.virt.hardware [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 527.418862] env[61987]: DEBUG nova.virt.hardware [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 527.419195] env[61987]: DEBUG nova.virt.hardware [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 527.419553] env[61987]: DEBUG nova.virt.hardware [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 527.419807] env[61987]: DEBUG nova.virt.hardware [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 527.420046] env[61987]: DEBUG nova.virt.hardware [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 527.420288] env[61987]: DEBUG nova.virt.hardware [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 527.420484] env[61987]: DEBUG nova.virt.hardware [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 527.421391] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6afff306-f4fb-4a4f-96e7-a88b38b6e378 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.429933] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de011e0-0ea7-41d6-956d-e3d1a322c50c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.628246] env[61987]: DEBUG nova.network.neutron [req-4fb3a036-ef55-4b8b-a69a-329afacc7bef req-cc56c294-cb18-4335-81f6-05d7c2901249 service nova] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 527.657392] env[61987]: DEBUG nova.network.neutron [-] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.734935] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.401s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.735353] env[61987]: DEBUG nova.compute.manager [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 527.738433] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.810s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.740034] env[61987]: INFO nova.compute.claims [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 527.907986] env[61987]: ERROR nova.compute.manager [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3f3ca432-e7a4-4a2f-9086-3983a2f72e29, please check neutron logs for more information. [ 527.907986] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 527.907986] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 527.907986] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 527.907986] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 527.907986] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 527.907986] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 527.907986] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 527.907986] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.907986] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 527.907986] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.907986] env[61987]: ERROR nova.compute.manager raise self.value [ 527.907986] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 527.907986] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 527.907986] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.907986] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 527.908614] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.908614] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 527.908614] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3f3ca432-e7a4-4a2f-9086-3983a2f72e29, please check neutron logs for more information. [ 527.908614] env[61987]: ERROR nova.compute.manager [ 527.908614] env[61987]: Traceback (most recent call last): [ 527.908614] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 527.908614] env[61987]: listener.cb(fileno) [ 527.908614] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 527.908614] env[61987]: result = function(*args, **kwargs) [ 527.908614] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 527.908614] env[61987]: return func(*args, **kwargs) [ 527.908614] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 527.908614] env[61987]: raise e [ 527.908614] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 527.908614] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 527.908614] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 527.908614] env[61987]: created_port_ids = self._update_ports_for_instance( [ 527.908614] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 527.908614] env[61987]: with excutils.save_and_reraise_exception(): [ 527.908614] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.908614] env[61987]: self.force_reraise() [ 527.908614] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.908614] env[61987]: raise self.value [ 527.908614] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 527.908614] env[61987]: updated_port = self._update_port( [ 527.908614] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.908614] env[61987]: _ensure_no_port_binding_failure(port) [ 527.908614] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.908614] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 527.909424] env[61987]: nova.exception.PortBindingFailed: Binding failed for port 3f3ca432-e7a4-4a2f-9086-3983a2f72e29, please check neutron logs for more information. [ 527.909424] env[61987]: Removing descriptor: 16 [ 527.909424] env[61987]: ERROR nova.compute.manager [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3f3ca432-e7a4-4a2f-9086-3983a2f72e29, please check neutron logs for more information. [ 527.909424] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Traceback (most recent call last): [ 527.909424] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 527.909424] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] yield resources [ 527.909424] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 527.909424] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] self.driver.spawn(context, instance, image_meta, [ 527.909424] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 527.909424] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] self._vmops.spawn(context, instance, image_meta, injected_files, [ 527.909424] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 527.909424] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] vm_ref = self.build_virtual_machine(instance, [ 527.909871] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 527.909871] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] vif_infos = vmwarevif.get_vif_info(self._session, [ 527.909871] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 527.909871] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] for vif in network_info: [ 527.909871] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 527.909871] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] return self._sync_wrapper(fn, *args, **kwargs) [ 527.909871] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 527.909871] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] self.wait() [ 527.909871] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 527.909871] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] self[:] = self._gt.wait() [ 527.909871] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 527.909871] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] return self._exit_event.wait() [ 527.909871] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 527.910251] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] result = hub.switch() [ 527.910251] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 527.910251] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] return self.greenlet.switch() [ 527.910251] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 527.910251] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] result = function(*args, **kwargs) [ 527.910251] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 527.910251] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] return func(*args, **kwargs) [ 527.910251] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 527.910251] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] raise e [ 527.910251] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 527.910251] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] nwinfo = self.network_api.allocate_for_instance( [ 527.910251] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 527.910251] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] created_port_ids = self._update_ports_for_instance( [ 527.910672] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 527.910672] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] with excutils.save_and_reraise_exception(): [ 527.910672] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.910672] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] self.force_reraise() [ 527.910672] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.910672] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] raise self.value [ 527.910672] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 527.910672] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] updated_port = self._update_port( [ 527.910672] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.910672] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] _ensure_no_port_binding_failure(port) [ 527.910672] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.910672] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] raise exception.PortBindingFailed(port_id=port['id']) [ 527.911021] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] nova.exception.PortBindingFailed: Binding failed for port 3f3ca432-e7a4-4a2f-9086-3983a2f72e29, please check neutron logs for more information. [ 527.911021] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] [ 527.911021] env[61987]: INFO nova.compute.manager [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Terminating instance [ 527.972763] env[61987]: DEBUG nova.network.neutron [req-4fb3a036-ef55-4b8b-a69a-329afacc7bef req-cc56c294-cb18-4335-81f6-05d7c2901249 service nova] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.159860] env[61987]: INFO nova.compute.manager [-] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Took 1.06 seconds to deallocate network for instance. [ 528.162170] env[61987]: DEBUG nova.compute.claims [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 528.162601] env[61987]: DEBUG oslo_concurrency.lockutils [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.243981] env[61987]: DEBUG nova.compute.utils [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 528.254124] env[61987]: DEBUG nova.compute.manager [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 528.254507] env[61987]: DEBUG nova.network.neutron [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 528.349961] env[61987]: DEBUG nova.policy [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bbf29f26e52443a18b62d8d805996952', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f72b5d838a1b4d9abbd298925f7b7508', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 528.419569] env[61987]: DEBUG oslo_concurrency.lockutils [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Acquiring lock "refresh_cache-25f011ba-eb0b-4295-be8a-d223584ada63" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 528.419569] env[61987]: DEBUG oslo_concurrency.lockutils [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Acquired lock "refresh_cache-25f011ba-eb0b-4295-be8a-d223584ada63" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 528.419723] env[61987]: DEBUG nova.network.neutron [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 528.479677] env[61987]: DEBUG oslo_concurrency.lockutils [req-4fb3a036-ef55-4b8b-a69a-329afacc7bef req-cc56c294-cb18-4335-81f6-05d7c2901249 service nova] Releasing lock "refresh_cache-31c175f9-20a5-4e38-ba90-972b847ab65e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.533467] env[61987]: DEBUG nova.compute.manager [req-cfd83555-b1ce-4c2b-b72c-ace77f720d73 req-a1577ac4-584b-4a7c-8bd9-d36b714c1790 service nova] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Received event network-changed-3f3ca432-e7a4-4a2f-9086-3983a2f72e29 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 528.533467] env[61987]: DEBUG nova.compute.manager [req-cfd83555-b1ce-4c2b-b72c-ace77f720d73 req-a1577ac4-584b-4a7c-8bd9-d36b714c1790 service nova] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Refreshing instance network info cache due to event network-changed-3f3ca432-e7a4-4a2f-9086-3983a2f72e29. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 528.533467] env[61987]: DEBUG oslo_concurrency.lockutils [req-cfd83555-b1ce-4c2b-b72c-ace77f720d73 req-a1577ac4-584b-4a7c-8bd9-d36b714c1790 service nova] Acquiring lock "refresh_cache-25f011ba-eb0b-4295-be8a-d223584ada63" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 528.759086] env[61987]: DEBUG nova.compute.manager [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 528.964500] env[61987]: DEBUG nova.network.neutron [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 528.973493] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26335be6-c9c2-4e2d-8722-20327532e0fa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.978872] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d631126-268e-461d-b19f-ed377172dc45 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.015308] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9e8ead1-596c-45e8-be19-a9022515ffd5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.023584] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-380085d3-6fcd-48fd-b4c9-ddc18c5fe1cb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.039629] env[61987]: DEBUG nova.compute.provider_tree [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 529.136730] env[61987]: DEBUG nova.network.neutron [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.236231] env[61987]: DEBUG nova.network.neutron [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Successfully created port: f80fb674-a7f0-4ea0-85b8-a7b404d94b8c {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 529.498493] env[61987]: DEBUG nova.compute.manager [req-eacedcc1-a88b-492e-bc1b-b7c7547dd044 req-746bffb2-4476-4dcf-ad3d-1f40f82eebe7 service nova] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Received event network-vif-deleted-691fbcaf-62ed-49e6-b5c0-06b24549eca4 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 529.542641] env[61987]: DEBUG nova.scheduler.client.report [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 529.639112] env[61987]: DEBUG oslo_concurrency.lockutils [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Releasing lock "refresh_cache-25f011ba-eb0b-4295-be8a-d223584ada63" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 529.639880] env[61987]: DEBUG nova.compute.manager [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 529.640158] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 529.640566] env[61987]: DEBUG oslo_concurrency.lockutils [req-cfd83555-b1ce-4c2b-b72c-ace77f720d73 req-a1577ac4-584b-4a7c-8bd9-d36b714c1790 service nova] Acquired lock "refresh_cache-25f011ba-eb0b-4295-be8a-d223584ada63" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 529.640705] env[61987]: DEBUG nova.network.neutron [req-cfd83555-b1ce-4c2b-b72c-ace77f720d73 req-a1577ac4-584b-4a7c-8bd9-d36b714c1790 service nova] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Refreshing network info cache for port 3f3ca432-e7a4-4a2f-9086-3983a2f72e29 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 529.642234] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dc66581e-6fd9-41ec-a54c-ed0d4b5f2592 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.657020] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb8e9d9c-799a-455a-9049-27e5e9a7efd3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.681008] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 25f011ba-eb0b-4295-be8a-d223584ada63 could not be found. [ 529.683542] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 529.683542] env[61987]: INFO nova.compute.manager [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Took 0.04 seconds to destroy the instance on the hypervisor. [ 529.683542] env[61987]: DEBUG oslo.service.loopingcall [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 529.683542] env[61987]: DEBUG nova.compute.manager [-] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 529.683542] env[61987]: DEBUG nova.network.neutron [-] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 529.721961] env[61987]: DEBUG nova.network.neutron [-] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 529.779621] env[61987]: DEBUG nova.compute.manager [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 529.814715] env[61987]: DEBUG nova.virt.hardware [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 529.815436] env[61987]: DEBUG nova.virt.hardware [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 529.815436] env[61987]: DEBUG nova.virt.hardware [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 529.815436] env[61987]: DEBUG nova.virt.hardware [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 529.815436] env[61987]: DEBUG nova.virt.hardware [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 529.815724] env[61987]: DEBUG nova.virt.hardware [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 529.815760] env[61987]: DEBUG nova.virt.hardware [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 529.816040] env[61987]: DEBUG nova.virt.hardware [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 529.822148] env[61987]: DEBUG nova.virt.hardware [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 529.822148] env[61987]: DEBUG nova.virt.hardware [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 529.822148] env[61987]: DEBUG nova.virt.hardware [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 529.822148] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9224fd4b-1770-48d0-9299-35867cb6df2b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.834670] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d61a281-b73f-4f29-b8ef-fa85cb6c4457 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.902751] env[61987]: ERROR nova.compute.manager [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cd2e6086-23ab-4064-a665-7e4b92b07e6f, please check neutron logs for more information. [ 529.902751] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 529.902751] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 529.902751] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 529.902751] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.902751] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 529.902751] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.902751] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 529.902751] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.902751] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 529.902751] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.902751] env[61987]: ERROR nova.compute.manager raise self.value [ 529.902751] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.902751] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 529.902751] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.902751] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 529.903506] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.903506] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 529.903506] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cd2e6086-23ab-4064-a665-7e4b92b07e6f, please check neutron logs for more information. [ 529.903506] env[61987]: ERROR nova.compute.manager [ 529.903506] env[61987]: Traceback (most recent call last): [ 529.903506] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 529.903506] env[61987]: listener.cb(fileno) [ 529.903506] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.903506] env[61987]: result = function(*args, **kwargs) [ 529.903506] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 529.903506] env[61987]: return func(*args, **kwargs) [ 529.903506] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 529.903506] env[61987]: raise e [ 529.903506] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 529.903506] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 529.903506] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.903506] env[61987]: created_port_ids = self._update_ports_for_instance( [ 529.903506] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.903506] env[61987]: with excutils.save_and_reraise_exception(): [ 529.903506] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.903506] env[61987]: self.force_reraise() [ 529.903506] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.903506] env[61987]: raise self.value [ 529.903506] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.903506] env[61987]: updated_port = self._update_port( [ 529.903506] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.903506] env[61987]: _ensure_no_port_binding_failure(port) [ 529.903506] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.903506] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 529.907550] env[61987]: nova.exception.PortBindingFailed: Binding failed for port cd2e6086-23ab-4064-a665-7e4b92b07e6f, please check neutron logs for more information. [ 529.907550] env[61987]: Removing descriptor: 15 [ 529.907550] env[61987]: ERROR nova.compute.manager [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cd2e6086-23ab-4064-a665-7e4b92b07e6f, please check neutron logs for more information. [ 529.907550] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Traceback (most recent call last): [ 529.907550] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 529.907550] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] yield resources [ 529.907550] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 529.907550] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] self.driver.spawn(context, instance, image_meta, [ 529.907550] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 529.907550] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] self._vmops.spawn(context, instance, image_meta, injected_files, [ 529.907550] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 529.907550] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] vm_ref = self.build_virtual_machine(instance, [ 529.908632] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 529.908632] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] vif_infos = vmwarevif.get_vif_info(self._session, [ 529.908632] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 529.908632] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] for vif in network_info: [ 529.908632] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 529.908632] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] return self._sync_wrapper(fn, *args, **kwargs) [ 529.908632] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 529.908632] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] self.wait() [ 529.908632] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 529.908632] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] self[:] = self._gt.wait() [ 529.908632] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 529.908632] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] return self._exit_event.wait() [ 529.908632] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 529.909104] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] result = hub.switch() [ 529.909104] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 529.909104] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] return self.greenlet.switch() [ 529.909104] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.909104] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] result = function(*args, **kwargs) [ 529.909104] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 529.909104] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] return func(*args, **kwargs) [ 529.909104] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 529.909104] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] raise e [ 529.909104] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 529.909104] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] nwinfo = self.network_api.allocate_for_instance( [ 529.909104] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 529.909104] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] created_port_ids = self._update_ports_for_instance( [ 529.910132] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 529.910132] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] with excutils.save_and_reraise_exception(): [ 529.910132] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.910132] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] self.force_reraise() [ 529.910132] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.910132] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] raise self.value [ 529.910132] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 529.910132] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] updated_port = self._update_port( [ 529.910132] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.910132] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] _ensure_no_port_binding_failure(port) [ 529.910132] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.910132] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] raise exception.PortBindingFailed(port_id=port['id']) [ 529.913197] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] nova.exception.PortBindingFailed: Binding failed for port cd2e6086-23ab-4064-a665-7e4b92b07e6f, please check neutron logs for more information. [ 529.913197] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] [ 529.913197] env[61987]: INFO nova.compute.manager [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Terminating instance [ 530.008210] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Acquiring lock "7a94f776-579d-41e8-94aa-6095b3fd69c6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.008210] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Lock "7a94f776-579d-41e8-94aa-6095b3fd69c6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.048994] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.310s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 530.049665] env[61987]: DEBUG nova.compute.manager [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 530.055349] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.901s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.060546] env[61987]: INFO nova.compute.claims [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 530.195784] env[61987]: DEBUG nova.network.neutron [req-cfd83555-b1ce-4c2b-b72c-ace77f720d73 req-a1577ac4-584b-4a7c-8bd9-d36b714c1790 service nova] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 530.228592] env[61987]: DEBUG nova.network.neutron [-] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.416589] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Acquiring lock "refresh_cache-7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 530.416589] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Acquired lock "refresh_cache-7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 530.416589] env[61987]: DEBUG nova.network.neutron [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 530.496038] env[61987]: DEBUG nova.network.neutron [req-cfd83555-b1ce-4c2b-b72c-ace77f720d73 req-a1577ac4-584b-4a7c-8bd9-d36b714c1790 service nova] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.565285] env[61987]: DEBUG nova.compute.utils [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 530.572986] env[61987]: DEBUG nova.compute.manager [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 530.572986] env[61987]: DEBUG nova.network.neutron [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 530.734089] env[61987]: INFO nova.compute.manager [-] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Took 1.05 seconds to deallocate network for instance. [ 530.739899] env[61987]: DEBUG nova.compute.claims [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 530.740094] env[61987]: DEBUG oslo_concurrency.lockutils [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.803334] env[61987]: DEBUG nova.policy [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd6a56eabf7cf4f6a97776274c2ff09c9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c3f08fe81311415ea6b115de508cb0b7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 530.978309] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 530.978309] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 530.978309] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Starting heal instance info cache {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10401}} [ 530.978309] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Rebuilding the list of instances to heal {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10405}} [ 530.998409] env[61987]: DEBUG nova.network.neutron [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 531.000562] env[61987]: DEBUG oslo_concurrency.lockutils [req-cfd83555-b1ce-4c2b-b72c-ace77f720d73 req-a1577ac4-584b-4a7c-8bd9-d36b714c1790 service nova] Releasing lock "refresh_cache-25f011ba-eb0b-4295-be8a-d223584ada63" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.077907] env[61987]: DEBUG nova.compute.manager [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 531.108396] env[61987]: DEBUG nova.compute.manager [req-b25bdeb8-bdee-4bb9-a404-797c882f070f req-4126a542-eb57-473c-905d-80bc86dd3d57 service nova] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Received event network-vif-deleted-3f3ca432-e7a4-4a2f-9086-3983a2f72e29 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 531.380616] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-179ddbb2-a9a6-463e-93fd-c25ce758dab0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.394335] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f44bdcb-e4a4-45ea-8e30-672fe53ce8ac {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.432937] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9edd6c24-27d2-450e-8e3f-7f3fde001d0c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.442819] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3b881a4-7635-4b95-aee5-4c2e25ee4ba1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.459977] env[61987]: DEBUG nova.compute.provider_tree [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 531.462629] env[61987]: DEBUG nova.network.neutron [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.483093] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Skipping network cache update for instance because it is Building. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10414}} [ 531.483420] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Skipping network cache update for instance because it is Building. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10414}} [ 531.483420] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Skipping network cache update for instance because it is Building. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10414}} [ 531.483496] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Skipping network cache update for instance because it is Building. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10414}} [ 531.484230] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Skipping network cache update for instance because it is Building. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10414}} [ 531.484230] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Skipping network cache update for instance because it is Building. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10414}} [ 531.484230] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Didn't find any instances for network info cache update. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10487}} [ 531.484440] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.484739] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.484898] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.485138] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.485346] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.485555] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.485728] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61987) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11020}} [ 531.485872] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 531.668910] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Acquiring lock "938dab0c-9b30-404b-bca2-a1280a252584" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.669227] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Lock "938dab0c-9b30-404b-bca2-a1280a252584" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.965645] env[61987]: DEBUG nova.scheduler.client.report [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 531.969696] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Releasing lock "refresh_cache-7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.972249] env[61987]: DEBUG nova.compute.manager [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 531.972249] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 531.972249] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-26e8d0c1-138f-4ff5-b0c5-0e29bd9d4298 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.982300] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdd24fdc-2533-486c-9d27-c1b3d13aa147 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.996033] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.011362] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac could not be found. [ 532.011362] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 532.011362] env[61987]: INFO nova.compute.manager [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Took 0.04 seconds to destroy the instance on the hypervisor. [ 532.011362] env[61987]: DEBUG oslo.service.loopingcall [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 532.011640] env[61987]: DEBUG nova.compute.manager [-] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 532.011640] env[61987]: DEBUG nova.network.neutron [-] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 532.085600] env[61987]: DEBUG nova.network.neutron [-] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 532.093072] env[61987]: DEBUG nova.compute.manager [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 532.128203] env[61987]: DEBUG nova.virt.hardware [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 532.128203] env[61987]: DEBUG nova.virt.hardware [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 532.128203] env[61987]: DEBUG nova.virt.hardware [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 532.128475] env[61987]: DEBUG nova.virt.hardware [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 532.128475] env[61987]: DEBUG nova.virt.hardware [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 532.128574] env[61987]: DEBUG nova.virt.hardware [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 532.128745] env[61987]: DEBUG nova.virt.hardware [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 532.128898] env[61987]: DEBUG nova.virt.hardware [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 532.135516] env[61987]: DEBUG nova.virt.hardware [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 532.135880] env[61987]: DEBUG nova.virt.hardware [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 532.136091] env[61987]: DEBUG nova.virt.hardware [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 532.137807] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f4cc50-92b8-4820-8ccd-070751cc60f4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.146505] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46974251-e854-4f16-bffc-f60cdaea96d8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.256211] env[61987]: DEBUG nova.compute.manager [req-a7ffe2b8-1ce8-4a73-a06b-614b85514dbd req-2b024fb9-8f2b-4297-9c26-62bb1469643f service nova] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Received event network-changed-cd2e6086-23ab-4064-a665-7e4b92b07e6f {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 532.256430] env[61987]: DEBUG nova.compute.manager [req-a7ffe2b8-1ce8-4a73-a06b-614b85514dbd req-2b024fb9-8f2b-4297-9c26-62bb1469643f service nova] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Refreshing instance network info cache due to event network-changed-cd2e6086-23ab-4064-a665-7e4b92b07e6f. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 532.256702] env[61987]: DEBUG oslo_concurrency.lockutils [req-a7ffe2b8-1ce8-4a73-a06b-614b85514dbd req-2b024fb9-8f2b-4297-9c26-62bb1469643f service nova] Acquiring lock "refresh_cache-7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 532.256907] env[61987]: DEBUG oslo_concurrency.lockutils [req-a7ffe2b8-1ce8-4a73-a06b-614b85514dbd req-2b024fb9-8f2b-4297-9c26-62bb1469643f service nova] Acquired lock "refresh_cache-7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 532.257060] env[61987]: DEBUG nova.network.neutron [req-a7ffe2b8-1ce8-4a73-a06b-614b85514dbd req-2b024fb9-8f2b-4297-9c26-62bb1469643f service nova] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Refreshing network info cache for port cd2e6086-23ab-4064-a665-7e4b92b07e6f {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 532.460946] env[61987]: DEBUG nova.network.neutron [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Successfully created port: 516b0723-ba96-4f21-9f78-8765da5d19a8 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 532.473954] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.418s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.474574] env[61987]: DEBUG nova.compute.manager [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 532.477891] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.943s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.479329] env[61987]: INFO nova.compute.claims [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 532.587534] env[61987]: DEBUG nova.network.neutron [-] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 532.818611] env[61987]: DEBUG nova.network.neutron [req-a7ffe2b8-1ce8-4a73-a06b-614b85514dbd req-2b024fb9-8f2b-4297-9c26-62bb1469643f service nova] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 532.984897] env[61987]: DEBUG nova.compute.utils [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 532.991043] env[61987]: DEBUG nova.compute.manager [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 532.991043] env[61987]: DEBUG nova.network.neutron [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 533.092089] env[61987]: INFO nova.compute.manager [-] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Took 1.08 seconds to deallocate network for instance. [ 533.097083] env[61987]: DEBUG nova.compute.claims [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 533.097467] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.148788] env[61987]: DEBUG nova.network.neutron [req-a7ffe2b8-1ce8-4a73-a06b-614b85514dbd req-2b024fb9-8f2b-4297-9c26-62bb1469643f service nova] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.219724] env[61987]: DEBUG nova.policy [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'df7c9201637a4ab5b7b6233e9edce96d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '421f400781744330acda20e70064029d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 533.420266] env[61987]: DEBUG oslo_concurrency.lockutils [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Acquiring lock "6bf0d1f1-d079-4956-bc9a-86ba66885f3a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.420364] env[61987]: DEBUG oslo_concurrency.lockutils [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Lock "6bf0d1f1-d079-4956-bc9a-86ba66885f3a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.492746] env[61987]: DEBUG nova.compute.manager [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 533.657037] env[61987]: DEBUG oslo_concurrency.lockutils [req-a7ffe2b8-1ce8-4a73-a06b-614b85514dbd req-2b024fb9-8f2b-4297-9c26-62bb1469643f service nova] Releasing lock "refresh_cache-7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.657758] env[61987]: DEBUG nova.compute.manager [req-a7ffe2b8-1ce8-4a73-a06b-614b85514dbd req-2b024fb9-8f2b-4297-9c26-62bb1469643f service nova] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Received event network-vif-deleted-cd2e6086-23ab-4064-a665-7e4b92b07e6f {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 533.766535] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64d08d32-1c3c-4b46-bf6d-5590880ffe98 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.775892] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f79306c-b7c7-4e06-8085-4b9ccd8a8aff {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.810600] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1df0416c-6b9a-4462-a683-41d5ae7a980d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.819150] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e28c2472-53e9-4dc4-a804-be3cc9579afb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.834142] env[61987]: DEBUG nova.compute.provider_tree [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 534.285152] env[61987]: DEBUG nova.network.neutron [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Successfully created port: 76fe4435-5b40-4bf3-82ac-e4268f7a3ddd {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 534.339165] env[61987]: DEBUG nova.scheduler.client.report [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 534.385473] env[61987]: ERROR nova.compute.manager [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f80fb674-a7f0-4ea0-85b8-a7b404d94b8c, please check neutron logs for more information. [ 534.385473] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 534.385473] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 534.385473] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 534.385473] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 534.385473] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 534.385473] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 534.385473] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 534.385473] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.385473] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 534.385473] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.385473] env[61987]: ERROR nova.compute.manager raise self.value [ 534.385473] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 534.385473] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 534.385473] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.385473] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 534.386567] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.386567] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 534.386567] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f80fb674-a7f0-4ea0-85b8-a7b404d94b8c, please check neutron logs for more information. [ 534.386567] env[61987]: ERROR nova.compute.manager [ 534.386567] env[61987]: Traceback (most recent call last): [ 534.386567] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 534.386567] env[61987]: listener.cb(fileno) [ 534.386567] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.386567] env[61987]: result = function(*args, **kwargs) [ 534.386567] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 534.386567] env[61987]: return func(*args, **kwargs) [ 534.386567] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 534.386567] env[61987]: raise e [ 534.386567] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 534.386567] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 534.386567] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 534.386567] env[61987]: created_port_ids = self._update_ports_for_instance( [ 534.386567] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 534.386567] env[61987]: with excutils.save_and_reraise_exception(): [ 534.386567] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.386567] env[61987]: self.force_reraise() [ 534.386567] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.386567] env[61987]: raise self.value [ 534.386567] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 534.386567] env[61987]: updated_port = self._update_port( [ 534.386567] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.386567] env[61987]: _ensure_no_port_binding_failure(port) [ 534.386567] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.386567] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 534.389059] env[61987]: nova.exception.PortBindingFailed: Binding failed for port f80fb674-a7f0-4ea0-85b8-a7b404d94b8c, please check neutron logs for more information. [ 534.389059] env[61987]: Removing descriptor: 17 [ 534.389059] env[61987]: ERROR nova.compute.manager [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f80fb674-a7f0-4ea0-85b8-a7b404d94b8c, please check neutron logs for more information. [ 534.389059] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Traceback (most recent call last): [ 534.389059] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 534.389059] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] yield resources [ 534.389059] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 534.389059] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] self.driver.spawn(context, instance, image_meta, [ 534.389059] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 534.389059] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 534.389059] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 534.389059] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] vm_ref = self.build_virtual_machine(instance, [ 534.389460] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 534.389460] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] vif_infos = vmwarevif.get_vif_info(self._session, [ 534.389460] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 534.389460] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] for vif in network_info: [ 534.389460] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 534.389460] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] return self._sync_wrapper(fn, *args, **kwargs) [ 534.389460] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 534.389460] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] self.wait() [ 534.389460] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 534.389460] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] self[:] = self._gt.wait() [ 534.389460] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 534.389460] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] return self._exit_event.wait() [ 534.389460] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 534.389796] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] result = hub.switch() [ 534.389796] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 534.389796] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] return self.greenlet.switch() [ 534.389796] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.389796] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] result = function(*args, **kwargs) [ 534.389796] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 534.389796] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] return func(*args, **kwargs) [ 534.389796] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 534.389796] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] raise e [ 534.389796] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 534.389796] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] nwinfo = self.network_api.allocate_for_instance( [ 534.389796] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 534.389796] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] created_port_ids = self._update_ports_for_instance( [ 534.390159] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 534.390159] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] with excutils.save_and_reraise_exception(): [ 534.390159] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.390159] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] self.force_reraise() [ 534.390159] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.390159] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] raise self.value [ 534.390159] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 534.390159] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] updated_port = self._update_port( [ 534.390159] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.390159] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] _ensure_no_port_binding_failure(port) [ 534.390159] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.390159] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] raise exception.PortBindingFailed(port_id=port['id']) [ 534.390472] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] nova.exception.PortBindingFailed: Binding failed for port f80fb674-a7f0-4ea0-85b8-a7b404d94b8c, please check neutron logs for more information. [ 534.390472] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] [ 534.390472] env[61987]: INFO nova.compute.manager [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Terminating instance [ 534.503801] env[61987]: DEBUG nova.compute.manager [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 534.540705] env[61987]: DEBUG nova.virt.hardware [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 534.540815] env[61987]: DEBUG nova.virt.hardware [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 534.541113] env[61987]: DEBUG nova.virt.hardware [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 534.541194] env[61987]: DEBUG nova.virt.hardware [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 534.542033] env[61987]: DEBUG nova.virt.hardware [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 534.542033] env[61987]: DEBUG nova.virt.hardware [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 534.542033] env[61987]: DEBUG nova.virt.hardware [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 534.542033] env[61987]: DEBUG nova.virt.hardware [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 534.542033] env[61987]: DEBUG nova.virt.hardware [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 534.548499] env[61987]: DEBUG nova.virt.hardware [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 534.548499] env[61987]: DEBUG nova.virt.hardware [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 534.548499] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-025cd860-0412-4836-a862-a80be3672842 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.559469] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d14041e-205d-4eb6-aab0-bb14faadfaeb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.785000] env[61987]: DEBUG nova.compute.manager [req-461f2910-9a45-4b5b-a8d9-1c2b8d1596e6 req-1f62d9bd-2e08-4dd0-ad47-8476f4432ac6 service nova] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Received event network-changed-f80fb674-a7f0-4ea0-85b8-a7b404d94b8c {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 534.785228] env[61987]: DEBUG nova.compute.manager [req-461f2910-9a45-4b5b-a8d9-1c2b8d1596e6 req-1f62d9bd-2e08-4dd0-ad47-8476f4432ac6 service nova] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Refreshing instance network info cache due to event network-changed-f80fb674-a7f0-4ea0-85b8-a7b404d94b8c. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 534.785445] env[61987]: DEBUG oslo_concurrency.lockutils [req-461f2910-9a45-4b5b-a8d9-1c2b8d1596e6 req-1f62d9bd-2e08-4dd0-ad47-8476f4432ac6 service nova] Acquiring lock "refresh_cache-ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.785588] env[61987]: DEBUG oslo_concurrency.lockutils [req-461f2910-9a45-4b5b-a8d9-1c2b8d1596e6 req-1f62d9bd-2e08-4dd0-ad47-8476f4432ac6 service nova] Acquired lock "refresh_cache-ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.785746] env[61987]: DEBUG nova.network.neutron [req-461f2910-9a45-4b5b-a8d9-1c2b8d1596e6 req-1f62d9bd-2e08-4dd0-ad47-8476f4432ac6 service nova] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Refreshing network info cache for port f80fb674-a7f0-4ea0-85b8-a7b404d94b8c {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 534.847773] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.370s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 534.850885] env[61987]: DEBUG nova.compute.manager [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 534.852585] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.446s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.854212] env[61987]: INFO nova.compute.claims [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 534.893689] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Acquiring lock "refresh_cache-ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.358311] env[61987]: DEBUG nova.compute.utils [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 535.367031] env[61987]: DEBUG nova.network.neutron [req-461f2910-9a45-4b5b-a8d9-1c2b8d1596e6 req-1f62d9bd-2e08-4dd0-ad47-8476f4432ac6 service nova] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 535.368106] env[61987]: DEBUG nova.compute.manager [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 535.368448] env[61987]: DEBUG nova.network.neutron [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 535.722910] env[61987]: DEBUG nova.policy [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3011f049be54b2cb8922d646b94a310', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd828262de7b14d2ab38193d7d34e8f7a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 535.877677] env[61987]: DEBUG nova.compute.manager [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 535.977884] env[61987]: DEBUG nova.network.neutron [req-461f2910-9a45-4b5b-a8d9-1c2b8d1596e6 req-1f62d9bd-2e08-4dd0-ad47-8476f4432ac6 service nova] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.188647] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-528ecda6-66d5-4932-a4fc-f692982a24bf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.197453] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-909fd00e-cf9d-45a1-984a-4ef1595302c6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.242941] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f0ee567-7e7a-4e06-9317-1e10fb3ba2d7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.251554] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63f29276-1387-46d4-aef1-227087a150a9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.265695] env[61987]: DEBUG nova.compute.provider_tree [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 536.484377] env[61987]: DEBUG oslo_concurrency.lockutils [req-461f2910-9a45-4b5b-a8d9-1c2b8d1596e6 req-1f62d9bd-2e08-4dd0-ad47-8476f4432ac6 service nova] Releasing lock "refresh_cache-ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.484982] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Acquired lock "refresh_cache-ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.484982] env[61987]: DEBUG nova.network.neutron [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 536.769222] env[61987]: DEBUG nova.scheduler.client.report [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 536.891031] env[61987]: DEBUG nova.compute.manager [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 536.920816] env[61987]: DEBUG nova.virt.hardware [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 536.921247] env[61987]: DEBUG nova.virt.hardware [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 536.921247] env[61987]: DEBUG nova.virt.hardware [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 536.921408] env[61987]: DEBUG nova.virt.hardware [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 536.921531] env[61987]: DEBUG nova.virt.hardware [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 536.921671] env[61987]: DEBUG nova.virt.hardware [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 536.922186] env[61987]: DEBUG nova.virt.hardware [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 536.922186] env[61987]: DEBUG nova.virt.hardware [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 536.922375] env[61987]: DEBUG nova.virt.hardware [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 536.922511] env[61987]: DEBUG nova.virt.hardware [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 536.922888] env[61987]: DEBUG nova.virt.hardware [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 536.924657] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e8048e-f42b-4f43-b0b8-7ff81bb53ca7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.933455] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e4db725-0fa0-485e-b173-52dde1ddd09b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.038549] env[61987]: DEBUG nova.network.neutron [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 537.184457] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Acquiring lock "7961bd0d-de24-4e13-94a3-79c48c593813" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.184457] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Lock "7961bd0d-de24-4e13-94a3-79c48c593813" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.219561] env[61987]: DEBUG nova.network.neutron [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Successfully created port: 555557bc-6f57-400e-8dbb-216d588433f1 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 537.280543] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.427s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.283076] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 537.287149] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.863s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.288746] env[61987]: INFO nova.compute.claims [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 537.510129] env[61987]: ERROR nova.compute.manager [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 516b0723-ba96-4f21-9f78-8765da5d19a8, please check neutron logs for more information. [ 537.510129] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 537.510129] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 537.510129] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 537.510129] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 537.510129] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 537.510129] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 537.510129] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 537.510129] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.510129] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 537.510129] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.510129] env[61987]: ERROR nova.compute.manager raise self.value [ 537.510129] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 537.510129] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 537.510129] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.510129] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 537.510592] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.510592] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 537.510592] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 516b0723-ba96-4f21-9f78-8765da5d19a8, please check neutron logs for more information. [ 537.510592] env[61987]: ERROR nova.compute.manager [ 537.510592] env[61987]: Traceback (most recent call last): [ 537.510592] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 537.510592] env[61987]: listener.cb(fileno) [ 537.510592] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.510592] env[61987]: result = function(*args, **kwargs) [ 537.510592] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 537.510592] env[61987]: return func(*args, **kwargs) [ 537.510592] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 537.510592] env[61987]: raise e [ 537.510592] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 537.510592] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 537.510592] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 537.510592] env[61987]: created_port_ids = self._update_ports_for_instance( [ 537.510592] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 537.510592] env[61987]: with excutils.save_and_reraise_exception(): [ 537.510592] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.510592] env[61987]: self.force_reraise() [ 537.510592] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.510592] env[61987]: raise self.value [ 537.510592] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 537.510592] env[61987]: updated_port = self._update_port( [ 537.510592] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.510592] env[61987]: _ensure_no_port_binding_failure(port) [ 537.510592] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.510592] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 537.511289] env[61987]: nova.exception.PortBindingFailed: Binding failed for port 516b0723-ba96-4f21-9f78-8765da5d19a8, please check neutron logs for more information. [ 537.511289] env[61987]: Removing descriptor: 15 [ 537.511289] env[61987]: ERROR nova.compute.manager [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 516b0723-ba96-4f21-9f78-8765da5d19a8, please check neutron logs for more information. [ 537.511289] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Traceback (most recent call last): [ 537.511289] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 537.511289] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] yield resources [ 537.511289] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 537.511289] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] self.driver.spawn(context, instance, image_meta, [ 537.511289] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 537.511289] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 537.511289] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 537.511289] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] vm_ref = self.build_virtual_machine(instance, [ 537.511593] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 537.511593] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] vif_infos = vmwarevif.get_vif_info(self._session, [ 537.511593] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 537.511593] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] for vif in network_info: [ 537.511593] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 537.511593] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] return self._sync_wrapper(fn, *args, **kwargs) [ 537.511593] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 537.511593] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] self.wait() [ 537.511593] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 537.511593] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] self[:] = self._gt.wait() [ 537.511593] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 537.511593] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] return self._exit_event.wait() [ 537.511593] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 537.511950] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] result = hub.switch() [ 537.511950] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 537.511950] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] return self.greenlet.switch() [ 537.511950] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.511950] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] result = function(*args, **kwargs) [ 537.511950] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 537.511950] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] return func(*args, **kwargs) [ 537.511950] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 537.511950] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] raise e [ 537.511950] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 537.511950] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] nwinfo = self.network_api.allocate_for_instance( [ 537.511950] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 537.511950] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] created_port_ids = self._update_ports_for_instance( [ 537.512274] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 537.512274] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] with excutils.save_and_reraise_exception(): [ 537.512274] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.512274] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] self.force_reraise() [ 537.512274] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.512274] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] raise self.value [ 537.512274] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 537.512274] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] updated_port = self._update_port( [ 537.512274] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.512274] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] _ensure_no_port_binding_failure(port) [ 537.512274] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.512274] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] raise exception.PortBindingFailed(port_id=port['id']) [ 537.512573] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] nova.exception.PortBindingFailed: Binding failed for port 516b0723-ba96-4f21-9f78-8765da5d19a8, please check neutron logs for more information. [ 537.512573] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] [ 537.512573] env[61987]: INFO nova.compute.manager [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Terminating instance [ 537.543776] env[61987]: DEBUG nova.compute.manager [req-7b479035-2cac-41d1-83df-18f2f8302074 req-2b97884f-8118-4ea3-ae0b-049c309bc589 service nova] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Received event network-changed-516b0723-ba96-4f21-9f78-8765da5d19a8 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 537.544116] env[61987]: DEBUG nova.compute.manager [req-7b479035-2cac-41d1-83df-18f2f8302074 req-2b97884f-8118-4ea3-ae0b-049c309bc589 service nova] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Refreshing instance network info cache due to event network-changed-516b0723-ba96-4f21-9f78-8765da5d19a8. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 537.544306] env[61987]: DEBUG oslo_concurrency.lockutils [req-7b479035-2cac-41d1-83df-18f2f8302074 req-2b97884f-8118-4ea3-ae0b-049c309bc589 service nova] Acquiring lock "refresh_cache-71316e2e-cda6-461c-8004-ff80fc2da90c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.544434] env[61987]: DEBUG oslo_concurrency.lockutils [req-7b479035-2cac-41d1-83df-18f2f8302074 req-2b97884f-8118-4ea3-ae0b-049c309bc589 service nova] Acquired lock "refresh_cache-71316e2e-cda6-461c-8004-ff80fc2da90c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.544599] env[61987]: DEBUG nova.network.neutron [req-7b479035-2cac-41d1-83df-18f2f8302074 req-2b97884f-8118-4ea3-ae0b-049c309bc589 service nova] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Refreshing network info cache for port 516b0723-ba96-4f21-9f78-8765da5d19a8 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 537.603962] env[61987]: DEBUG nova.network.neutron [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.796134] env[61987]: DEBUG nova.compute.utils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 537.796134] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 537.797556] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 537.819827] env[61987]: DEBUG nova.compute.manager [req-14aa1b54-7e7c-434d-b25a-3f2d1f3833ef req-ffe997b8-98f3-4ba0-a25f-567557d11198 service nova] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Received event network-vif-deleted-f80fb674-a7f0-4ea0-85b8-a7b404d94b8c {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 537.996210] env[61987]: DEBUG nova.policy [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1d3a436483c42f982ff9583ba07f9cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2262a448840c4564b229301f2ef9ec43', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 538.017141] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Acquiring lock "refresh_cache-71316e2e-cda6-461c-8004-ff80fc2da90c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.107180] env[61987]: DEBUG nova.network.neutron [req-7b479035-2cac-41d1-83df-18f2f8302074 req-2b97884f-8118-4ea3-ae0b-049c309bc589 service nova] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 538.108705] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Releasing lock "refresh_cache-ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.109583] env[61987]: DEBUG nova.compute.manager [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 538.116000] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 538.116000] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-52c3ef2c-65c0-48aa-b9f4-de08c56e5daa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.125457] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-805f756d-576b-40a9-8c39-ea79ba49da63 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.147514] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6 could not be found. [ 538.147756] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 538.147896] env[61987]: INFO nova.compute.manager [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Took 0.03 seconds to destroy the instance on the hypervisor. [ 538.148151] env[61987]: DEBUG oslo.service.loopingcall [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 538.148357] env[61987]: DEBUG nova.compute.manager [-] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 538.148450] env[61987]: DEBUG nova.network.neutron [-] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 538.212970] env[61987]: DEBUG nova.network.neutron [-] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 538.247918] env[61987]: DEBUG nova.network.neutron [req-7b479035-2cac-41d1-83df-18f2f8302074 req-2b97884f-8118-4ea3-ae0b-049c309bc589 service nova] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.307144] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 538.570418] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-789f1443-e718-4ca3-9991-699221d522d5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.579540] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-544015d0-b2f3-4c33-b2db-a29140b7dffe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.614922] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d404bb-7035-4f89-91db-e05016fe7254 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.622586] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-911f73fd-1f42-485a-8258-9725d6af98b1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.637374] env[61987]: DEBUG nova.compute.provider_tree [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 538.716209] env[61987]: DEBUG nova.network.neutron [-] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.752685] env[61987]: DEBUG oslo_concurrency.lockutils [req-7b479035-2cac-41d1-83df-18f2f8302074 req-2b97884f-8118-4ea3-ae0b-049c309bc589 service nova] Releasing lock "refresh_cache-71316e2e-cda6-461c-8004-ff80fc2da90c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.753210] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Acquired lock "refresh_cache-71316e2e-cda6-461c-8004-ff80fc2da90c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.753397] env[61987]: DEBUG nova.network.neutron [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 539.143101] env[61987]: DEBUG nova.scheduler.client.report [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 539.221530] env[61987]: INFO nova.compute.manager [-] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Took 1.07 seconds to deallocate network for instance. [ 539.224872] env[61987]: DEBUG nova.compute.claims [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 539.224872] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.292645] env[61987]: DEBUG nova.network.neutron [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 539.318878] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 539.349589] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 539.349903] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 539.350061] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 539.350206] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 539.350349] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 539.350498] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 539.350705] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 539.350879] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 539.351326] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 539.351618] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 539.351897] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 539.353959] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f00d5fb5-3348-4eb4-a256-e8272ce68494 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.364156] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cec8616f-56c2-4414-bca9-77f46fe6d058 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.472039] env[61987]: DEBUG nova.network.neutron [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.648018] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.359s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.648018] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 539.650181] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.182s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.651615] env[61987]: INFO nova.compute.claims [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 539.916751] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Successfully created port: 5de8b27a-666f-4c53-9ef8-4b92e5c5a592 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 539.974694] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Releasing lock "refresh_cache-71316e2e-cda6-461c-8004-ff80fc2da90c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.975189] env[61987]: DEBUG nova.compute.manager [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 539.975330] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 539.976051] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-43708488-953c-47ca-9b38-0ce135afeb18 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.986142] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d884cf23-bc60-41f8-9d09-5cdd5e5dfbfc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.008375] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 71316e2e-cda6-461c-8004-ff80fc2da90c could not be found. [ 540.008853] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 540.008853] env[61987]: INFO nova.compute.manager [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 540.009028] env[61987]: DEBUG oslo.service.loopingcall [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 540.009236] env[61987]: DEBUG nova.compute.manager [-] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 540.009324] env[61987]: DEBUG nova.network.neutron [-] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 540.044819] env[61987]: DEBUG nova.network.neutron [-] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 540.152154] env[61987]: DEBUG nova.compute.utils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 540.156239] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 540.156423] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 540.333739] env[61987]: DEBUG nova.policy [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1d3a436483c42f982ff9583ba07f9cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2262a448840c4564b229301f2ef9ec43', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 540.549686] env[61987]: DEBUG nova.network.neutron [-] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.665251] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 540.743297] env[61987]: DEBUG nova.compute.manager [req-c3bf27be-7887-4032-8abb-22315d05ee5e req-8cdfdcdc-8bca-47d7-98cf-bbd1429135f0 service nova] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Received event network-vif-deleted-516b0723-ba96-4f21-9f78-8765da5d19a8 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 540.769797] env[61987]: ERROR nova.compute.manager [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 76fe4435-5b40-4bf3-82ac-e4268f7a3ddd, please check neutron logs for more information. [ 540.769797] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 540.769797] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 540.769797] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 540.769797] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.769797] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 540.769797] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.769797] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 540.769797] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.769797] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 540.769797] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.769797] env[61987]: ERROR nova.compute.manager raise self.value [ 540.769797] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.769797] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 540.769797] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.769797] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 540.770724] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.770724] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 540.770724] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 76fe4435-5b40-4bf3-82ac-e4268f7a3ddd, please check neutron logs for more information. [ 540.770724] env[61987]: ERROR nova.compute.manager [ 540.770724] env[61987]: Traceback (most recent call last): [ 540.770724] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 540.770724] env[61987]: listener.cb(fileno) [ 540.770724] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.770724] env[61987]: result = function(*args, **kwargs) [ 540.770724] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 540.770724] env[61987]: return func(*args, **kwargs) [ 540.770724] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 540.770724] env[61987]: raise e [ 540.770724] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 540.770724] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 540.770724] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.770724] env[61987]: created_port_ids = self._update_ports_for_instance( [ 540.770724] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.770724] env[61987]: with excutils.save_and_reraise_exception(): [ 540.770724] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.770724] env[61987]: self.force_reraise() [ 540.770724] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.770724] env[61987]: raise self.value [ 540.770724] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.770724] env[61987]: updated_port = self._update_port( [ 540.770724] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.770724] env[61987]: _ensure_no_port_binding_failure(port) [ 540.770724] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.770724] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 540.772084] env[61987]: nova.exception.PortBindingFailed: Binding failed for port 76fe4435-5b40-4bf3-82ac-e4268f7a3ddd, please check neutron logs for more information. [ 540.772084] env[61987]: Removing descriptor: 18 [ 540.772084] env[61987]: ERROR nova.compute.manager [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 76fe4435-5b40-4bf3-82ac-e4268f7a3ddd, please check neutron logs for more information. [ 540.772084] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Traceback (most recent call last): [ 540.772084] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 540.772084] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] yield resources [ 540.772084] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 540.772084] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] self.driver.spawn(context, instance, image_meta, [ 540.772084] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 540.772084] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 540.772084] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 540.772084] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] vm_ref = self.build_virtual_machine(instance, [ 540.772661] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 540.772661] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] vif_infos = vmwarevif.get_vif_info(self._session, [ 540.772661] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 540.772661] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] for vif in network_info: [ 540.772661] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 540.772661] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] return self._sync_wrapper(fn, *args, **kwargs) [ 540.772661] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 540.772661] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] self.wait() [ 540.772661] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 540.772661] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] self[:] = self._gt.wait() [ 540.772661] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 540.772661] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] return self._exit_event.wait() [ 540.772661] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 540.774979] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] result = hub.switch() [ 540.774979] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 540.774979] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] return self.greenlet.switch() [ 540.774979] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.774979] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] result = function(*args, **kwargs) [ 540.774979] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 540.774979] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] return func(*args, **kwargs) [ 540.774979] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 540.774979] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] raise e [ 540.774979] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 540.774979] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] nwinfo = self.network_api.allocate_for_instance( [ 540.774979] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.774979] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] created_port_ids = self._update_ports_for_instance( [ 540.775671] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.775671] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] with excutils.save_and_reraise_exception(): [ 540.775671] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.775671] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] self.force_reraise() [ 540.775671] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.775671] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] raise self.value [ 540.775671] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.775671] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] updated_port = self._update_port( [ 540.775671] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.775671] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] _ensure_no_port_binding_failure(port) [ 540.775671] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.775671] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] raise exception.PortBindingFailed(port_id=port['id']) [ 540.776504] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] nova.exception.PortBindingFailed: Binding failed for port 76fe4435-5b40-4bf3-82ac-e4268f7a3ddd, please check neutron logs for more information. [ 540.776504] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] [ 540.776504] env[61987]: INFO nova.compute.manager [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Terminating instance [ 541.034466] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa9d669a-10cf-4101-a7d4-ffa87a524fe2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.044610] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c065ca-c320-4171-af9a-05cee7f56855 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.079927] env[61987]: INFO nova.compute.manager [-] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Took 1.07 seconds to deallocate network for instance. [ 541.081352] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08152fb1-c949-4b3b-b8d5-e84ae148e607 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.085324] env[61987]: DEBUG nova.compute.claims [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 541.085324] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.090602] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb15af08-cb3c-4314-8d21-26d92a6f8e95 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.105234] env[61987]: DEBUG nova.compute.provider_tree [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 541.278932] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Acquiring lock "refresh_cache-a0a004c3-6254-4f47-a080-2b9cc85a2a1d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.279219] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Acquired lock "refresh_cache-a0a004c3-6254-4f47-a080-2b9cc85a2a1d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.279453] env[61987]: DEBUG nova.network.neutron [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 541.504184] env[61987]: DEBUG nova.compute.manager [req-092d406b-7193-47e4-be4e-e0b733c81db5 req-d0d586d1-9925-4504-818f-3ec5e034df66 service nova] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Received event network-changed-76fe4435-5b40-4bf3-82ac-e4268f7a3ddd {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 541.504372] env[61987]: DEBUG nova.compute.manager [req-092d406b-7193-47e4-be4e-e0b733c81db5 req-d0d586d1-9925-4504-818f-3ec5e034df66 service nova] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Refreshing instance network info cache due to event network-changed-76fe4435-5b40-4bf3-82ac-e4268f7a3ddd. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 541.504556] env[61987]: DEBUG oslo_concurrency.lockutils [req-092d406b-7193-47e4-be4e-e0b733c81db5 req-d0d586d1-9925-4504-818f-3ec5e034df66 service nova] Acquiring lock "refresh_cache-a0a004c3-6254-4f47-a080-2b9cc85a2a1d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.610684] env[61987]: DEBUG nova.scheduler.client.report [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 541.683392] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 541.714177] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 541.714427] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 541.714803] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 541.714803] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 541.714945] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 541.715511] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 541.715909] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 541.716347] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 541.716918] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 541.716918] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 541.716918] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 541.719212] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb671053-35f2-4258-8d98-d0a02a186f0a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.728724] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-169e08c9-fc48-4762-8df4-c3d1d7296284 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.786570] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Successfully created port: 210424e1-1a59-4cf6-b39d-6ada30cc40e7 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 541.868206] env[61987]: DEBUG nova.network.neutron [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 542.117121] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.466s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 542.117711] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 542.124229] env[61987]: DEBUG oslo_concurrency.lockutils [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.962s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.273549] env[61987]: DEBUG nova.network.neutron [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.628024] env[61987]: DEBUG nova.compute.utils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 542.628024] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 542.628024] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 542.757443] env[61987]: DEBUG nova.policy [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b1d3a436483c42f982ff9583ba07f9cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2262a448840c4564b229301f2ef9ec43', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 542.782494] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Releasing lock "refresh_cache-a0a004c3-6254-4f47-a080-2b9cc85a2a1d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.782494] env[61987]: DEBUG nova.compute.manager [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 542.782662] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 542.782990] env[61987]: DEBUG oslo_concurrency.lockutils [req-092d406b-7193-47e4-be4e-e0b733c81db5 req-d0d586d1-9925-4504-818f-3ec5e034df66 service nova] Acquired lock "refresh_cache-a0a004c3-6254-4f47-a080-2b9cc85a2a1d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.783193] env[61987]: DEBUG nova.network.neutron [req-092d406b-7193-47e4-be4e-e0b733c81db5 req-d0d586d1-9925-4504-818f-3ec5e034df66 service nova] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Refreshing network info cache for port 76fe4435-5b40-4bf3-82ac-e4268f7a3ddd {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 542.785755] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-39db8826-e135-43a2-a244-be6d4b3a8180 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.799978] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9faf7db9-8c97-4e9b-9dda-5650817ddf04 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.830555] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a0a004c3-6254-4f47-a080-2b9cc85a2a1d could not be found. [ 542.832893] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 542.832893] env[61987]: INFO nova.compute.manager [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 542.832893] env[61987]: DEBUG oslo.service.loopingcall [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 542.832893] env[61987]: DEBUG nova.compute.manager [-] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 542.832893] env[61987]: DEBUG nova.network.neutron [-] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 542.959475] env[61987]: DEBUG nova.network.neutron [-] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 542.980339] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18555c6-55ab-428b-81b5-56bc15238070 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.990974] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aca9162-2fb5-4464-9c57-4e15278b4294 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.041216] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f2f7223-d9ed-4f22-b268-67ce881703a6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.049816] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ca28ef-c8b2-4c20-941f-332007bf7c75 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.066258] env[61987]: DEBUG nova.compute.provider_tree [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 543.132975] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 543.427772] env[61987]: DEBUG nova.network.neutron [req-092d406b-7193-47e4-be4e-e0b733c81db5 req-d0d586d1-9925-4504-818f-3ec5e034df66 service nova] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 543.467193] env[61987]: DEBUG nova.network.neutron [-] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.569562] env[61987]: DEBUG nova.scheduler.client.report [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 543.910034] env[61987]: DEBUG nova.network.neutron [req-092d406b-7193-47e4-be4e-e0b733c81db5 req-d0d586d1-9925-4504-818f-3ec5e034df66 service nova] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.973677] env[61987]: INFO nova.compute.manager [-] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Took 1.14 seconds to deallocate network for instance. [ 543.978191] env[61987]: DEBUG nova.compute.claims [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 543.978386] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.079372] env[61987]: DEBUG oslo_concurrency.lockutils [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.955s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.080015] env[61987]: ERROR nova.compute.manager [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 691fbcaf-62ed-49e6-b5c0-06b24549eca4, please check neutron logs for more information. [ 544.080015] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Traceback (most recent call last): [ 544.080015] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 544.080015] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] self.driver.spawn(context, instance, image_meta, [ 544.080015] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 544.080015] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 544.080015] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 544.080015] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] vm_ref = self.build_virtual_machine(instance, [ 544.080015] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 544.080015] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] vif_infos = vmwarevif.get_vif_info(self._session, [ 544.080015] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 544.080373] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] for vif in network_info: [ 544.080373] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 544.080373] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] return self._sync_wrapper(fn, *args, **kwargs) [ 544.080373] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 544.080373] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] self.wait() [ 544.080373] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 544.080373] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] self[:] = self._gt.wait() [ 544.080373] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 544.080373] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] return self._exit_event.wait() [ 544.080373] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 544.080373] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] result = hub.switch() [ 544.080373] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 544.080373] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] return self.greenlet.switch() [ 544.080718] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.080718] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] result = function(*args, **kwargs) [ 544.080718] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.080718] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] return func(*args, **kwargs) [ 544.080718] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 544.080718] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] raise e [ 544.080718] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 544.080718] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] nwinfo = self.network_api.allocate_for_instance( [ 544.080718] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.080718] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] created_port_ids = self._update_ports_for_instance( [ 544.080718] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.080718] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] with excutils.save_and_reraise_exception(): [ 544.080718] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.081054] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] self.force_reraise() [ 544.081054] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.081054] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] raise self.value [ 544.081054] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.081054] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] updated_port = self._update_port( [ 544.081054] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.081054] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] _ensure_no_port_binding_failure(port) [ 544.081054] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.081054] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] raise exception.PortBindingFailed(port_id=port['id']) [ 544.081054] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] nova.exception.PortBindingFailed: Binding failed for port 691fbcaf-62ed-49e6-b5c0-06b24549eca4, please check neutron logs for more information. [ 544.081054] env[61987]: ERROR nova.compute.manager [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] [ 544.081360] env[61987]: DEBUG nova.compute.utils [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Binding failed for port 691fbcaf-62ed-49e6-b5c0-06b24549eca4, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 544.082092] env[61987]: DEBUG oslo_concurrency.lockutils [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.342s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.100496] env[61987]: DEBUG nova.compute.manager [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Build of instance 31c175f9-20a5-4e38-ba90-972b847ab65e was re-scheduled: Binding failed for port 691fbcaf-62ed-49e6-b5c0-06b24549eca4, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 544.101026] env[61987]: DEBUG nova.compute.manager [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 544.101273] env[61987]: DEBUG oslo_concurrency.lockutils [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "refresh_cache-31c175f9-20a5-4e38-ba90-972b847ab65e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.101408] env[61987]: DEBUG oslo_concurrency.lockutils [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired lock "refresh_cache-31c175f9-20a5-4e38-ba90-972b847ab65e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.101594] env[61987]: DEBUG nova.network.neutron [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 544.108965] env[61987]: ERROR nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 210424e1-1a59-4cf6-b39d-6ada30cc40e7, please check neutron logs for more information. [ 544.108965] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 544.108965] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 544.108965] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 544.108965] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.108965] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 544.108965] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.108965] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 544.108965] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.108965] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 544.108965] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.108965] env[61987]: ERROR nova.compute.manager raise self.value [ 544.108965] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.108965] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 544.108965] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.108965] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 544.109510] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.109510] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 544.109510] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 210424e1-1a59-4cf6-b39d-6ada30cc40e7, please check neutron logs for more information. [ 544.109510] env[61987]: ERROR nova.compute.manager [ 544.109510] env[61987]: Traceback (most recent call last): [ 544.109510] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 544.109510] env[61987]: listener.cb(fileno) [ 544.109510] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.109510] env[61987]: result = function(*args, **kwargs) [ 544.109510] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.109510] env[61987]: return func(*args, **kwargs) [ 544.109510] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 544.109510] env[61987]: raise e [ 544.109510] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 544.109510] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 544.109510] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.109510] env[61987]: created_port_ids = self._update_ports_for_instance( [ 544.109510] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.109510] env[61987]: with excutils.save_and_reraise_exception(): [ 544.109510] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.109510] env[61987]: self.force_reraise() [ 544.109510] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.109510] env[61987]: raise self.value [ 544.109510] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.109510] env[61987]: updated_port = self._update_port( [ 544.109510] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.109510] env[61987]: _ensure_no_port_binding_failure(port) [ 544.109510] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.109510] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 544.111354] env[61987]: nova.exception.PortBindingFailed: Binding failed for port 210424e1-1a59-4cf6-b39d-6ada30cc40e7, please check neutron logs for more information. [ 544.111354] env[61987]: Removing descriptor: 15 [ 544.111354] env[61987]: ERROR nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 210424e1-1a59-4cf6-b39d-6ada30cc40e7, please check neutron logs for more information. [ 544.111354] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Traceback (most recent call last): [ 544.111354] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 544.111354] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] yield resources [ 544.111354] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 544.111354] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] self.driver.spawn(context, instance, image_meta, [ 544.111354] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 544.111354] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 544.111354] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 544.111354] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] vm_ref = self.build_virtual_machine(instance, [ 544.111669] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 544.111669] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] vif_infos = vmwarevif.get_vif_info(self._session, [ 544.111669] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 544.111669] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] for vif in network_info: [ 544.111669] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 544.111669] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] return self._sync_wrapper(fn, *args, **kwargs) [ 544.111669] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 544.111669] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] self.wait() [ 544.111669] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 544.111669] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] self[:] = self._gt.wait() [ 544.111669] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 544.111669] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] return self._exit_event.wait() [ 544.111669] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 544.114896] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] result = hub.switch() [ 544.114896] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 544.114896] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] return self.greenlet.switch() [ 544.114896] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.114896] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] result = function(*args, **kwargs) [ 544.114896] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.114896] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] return func(*args, **kwargs) [ 544.114896] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 544.114896] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] raise e [ 544.114896] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 544.114896] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] nwinfo = self.network_api.allocate_for_instance( [ 544.114896] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.114896] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] created_port_ids = self._update_ports_for_instance( [ 544.115719] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.115719] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] with excutils.save_and_reraise_exception(): [ 544.115719] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.115719] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] self.force_reraise() [ 544.115719] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.115719] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] raise self.value [ 544.115719] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.115719] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] updated_port = self._update_port( [ 544.115719] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.115719] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] _ensure_no_port_binding_failure(port) [ 544.115719] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.115719] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] raise exception.PortBindingFailed(port_id=port['id']) [ 544.116144] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] nova.exception.PortBindingFailed: Binding failed for port 210424e1-1a59-4cf6-b39d-6ada30cc40e7, please check neutron logs for more information. [ 544.116144] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] [ 544.116144] env[61987]: INFO nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Terminating instance [ 544.144133] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 544.184280] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 544.184570] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 544.184729] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 544.184908] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 544.185323] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 544.185756] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 544.186023] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 544.186193] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 544.186365] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 544.186710] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 544.186906] env[61987]: DEBUG nova.virt.hardware [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 544.187840] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6cd9c41-2c93-4587-80da-92429643baa8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.200518] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37764b3d-7bb5-4168-a236-375d01ff0744 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.327573] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Successfully created port: 30ce6103-be54-438d-930c-51e75dbcf857 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 544.379215] env[61987]: ERROR nova.compute.manager [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 555557bc-6f57-400e-8dbb-216d588433f1, please check neutron logs for more information. [ 544.379215] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 544.379215] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 544.379215] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 544.379215] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.379215] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 544.379215] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.379215] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 544.379215] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.379215] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 544.379215] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.379215] env[61987]: ERROR nova.compute.manager raise self.value [ 544.379215] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.379215] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 544.379215] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.379215] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 544.379723] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.379723] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 544.379723] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 555557bc-6f57-400e-8dbb-216d588433f1, please check neutron logs for more information. [ 544.379723] env[61987]: ERROR nova.compute.manager [ 544.381201] env[61987]: Traceback (most recent call last): [ 544.381201] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 544.381201] env[61987]: listener.cb(fileno) [ 544.381201] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.381201] env[61987]: result = function(*args, **kwargs) [ 544.381201] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.381201] env[61987]: return func(*args, **kwargs) [ 544.381201] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 544.381201] env[61987]: raise e [ 544.381201] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 544.381201] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 544.381201] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.381201] env[61987]: created_port_ids = self._update_ports_for_instance( [ 544.381201] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.381201] env[61987]: with excutils.save_and_reraise_exception(): [ 544.381201] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.381201] env[61987]: self.force_reraise() [ 544.381201] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.381201] env[61987]: raise self.value [ 544.381201] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.381201] env[61987]: updated_port = self._update_port( [ 544.381201] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.381201] env[61987]: _ensure_no_port_binding_failure(port) [ 544.381201] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.381201] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 544.381201] env[61987]: nova.exception.PortBindingFailed: Binding failed for port 555557bc-6f57-400e-8dbb-216d588433f1, please check neutron logs for more information. [ 544.381201] env[61987]: Removing descriptor: 16 [ 544.382729] env[61987]: ERROR nova.compute.manager [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 555557bc-6f57-400e-8dbb-216d588433f1, please check neutron logs for more information. [ 544.382729] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Traceback (most recent call last): [ 544.382729] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 544.382729] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] yield resources [ 544.382729] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 544.382729] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] self.driver.spawn(context, instance, image_meta, [ 544.382729] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 544.382729] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 544.382729] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 544.382729] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] vm_ref = self.build_virtual_machine(instance, [ 544.382729] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 544.383118] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] vif_infos = vmwarevif.get_vif_info(self._session, [ 544.383118] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 544.383118] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] for vif in network_info: [ 544.383118] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 544.383118] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] return self._sync_wrapper(fn, *args, **kwargs) [ 544.383118] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 544.383118] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] self.wait() [ 544.383118] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 544.383118] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] self[:] = self._gt.wait() [ 544.383118] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 544.383118] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] return self._exit_event.wait() [ 544.383118] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 544.383118] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] result = hub.switch() [ 544.383537] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 544.383537] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] return self.greenlet.switch() [ 544.383537] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.383537] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] result = function(*args, **kwargs) [ 544.383537] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.383537] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] return func(*args, **kwargs) [ 544.383537] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 544.383537] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] raise e [ 544.383537] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 544.383537] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] nwinfo = self.network_api.allocate_for_instance( [ 544.383537] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.383537] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] created_port_ids = self._update_ports_for_instance( [ 544.383537] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.383910] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] with excutils.save_and_reraise_exception(): [ 544.383910] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.383910] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] self.force_reraise() [ 544.383910] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.383910] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] raise self.value [ 544.383910] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.383910] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] updated_port = self._update_port( [ 544.383910] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.383910] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] _ensure_no_port_binding_failure(port) [ 544.383910] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.383910] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] raise exception.PortBindingFailed(port_id=port['id']) [ 544.383910] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] nova.exception.PortBindingFailed: Binding failed for port 555557bc-6f57-400e-8dbb-216d588433f1, please check neutron logs for more information. [ 544.383910] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] [ 544.384451] env[61987]: INFO nova.compute.manager [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Terminating instance [ 544.416189] env[61987]: DEBUG oslo_concurrency.lockutils [req-092d406b-7193-47e4-be4e-e0b733c81db5 req-d0d586d1-9925-4504-818f-3ec5e034df66 service nova] Releasing lock "refresh_cache-a0a004c3-6254-4f47-a080-2b9cc85a2a1d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.619975] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquiring lock "refresh_cache-1a9030ce-3a35-4760-96f4-4288c1d2f9d1" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.620571] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquired lock "refresh_cache-1a9030ce-3a35-4760-96f4-4288c1d2f9d1" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.620571] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 544.667956] env[61987]: DEBUG nova.network.neutron [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 544.891661] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Acquiring lock "bac5a3bc-5a58-4c1d-abb8-1aaf6151358a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.891889] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Lock "bac5a3bc-5a58-4c1d-abb8-1aaf6151358a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.892269] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "refresh_cache-9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.892467] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquired lock "refresh_cache-9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.892639] env[61987]: DEBUG nova.network.neutron [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 544.914242] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cfd8c74-12f6-4d44-aad9-f5a0f6ec52b9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.922523] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9f7b9a6-2063-4073-8bf1-1b5526226870 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.952654] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8259e489-9c73-41f0-85e7-8eeab5391c0a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.960101] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edbc6c13-a4e9-4d5a-b93d-dba292a81856 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.974419] env[61987]: DEBUG nova.compute.provider_tree [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 544.976375] env[61987]: DEBUG nova.network.neutron [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.135625] env[61987]: DEBUG nova.compute.manager [req-53a20691-729d-43ae-bf0a-763b9cbf21d9 req-155f83e5-c508-4850-9b19-fa5c217a50d3 service nova] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Received event network-vif-deleted-76fe4435-5b40-4bf3-82ac-e4268f7a3ddd {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 545.135935] env[61987]: DEBUG nova.compute.manager [req-53a20691-729d-43ae-bf0a-763b9cbf21d9 req-155f83e5-c508-4850-9b19-fa5c217a50d3 service nova] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Received event network-changed-555557bc-6f57-400e-8dbb-216d588433f1 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 545.135935] env[61987]: DEBUG nova.compute.manager [req-53a20691-729d-43ae-bf0a-763b9cbf21d9 req-155f83e5-c508-4850-9b19-fa5c217a50d3 service nova] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Refreshing instance network info cache due to event network-changed-555557bc-6f57-400e-8dbb-216d588433f1. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 545.136226] env[61987]: DEBUG oslo_concurrency.lockutils [req-53a20691-729d-43ae-bf0a-763b9cbf21d9 req-155f83e5-c508-4850-9b19-fa5c217a50d3 service nova] Acquiring lock "refresh_cache-9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.156949] env[61987]: DEBUG nova.compute.manager [req-7827b610-e0d2-4f78-8be3-6fc97638aefc req-cf652f89-d431-4fbd-a106-bfd03c2ce98d service nova] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Received event network-changed-210424e1-1a59-4cf6-b39d-6ada30cc40e7 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 545.158040] env[61987]: DEBUG nova.compute.manager [req-7827b610-e0d2-4f78-8be3-6fc97638aefc req-cf652f89-d431-4fbd-a106-bfd03c2ce98d service nova] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Refreshing instance network info cache due to event network-changed-210424e1-1a59-4cf6-b39d-6ada30cc40e7. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 545.158040] env[61987]: DEBUG oslo_concurrency.lockutils [req-7827b610-e0d2-4f78-8be3-6fc97638aefc req-cf652f89-d431-4fbd-a106-bfd03c2ce98d service nova] Acquiring lock "refresh_cache-1a9030ce-3a35-4760-96f4-4288c1d2f9d1" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.161453] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 545.430121] env[61987]: DEBUG nova.network.neutron [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 545.480569] env[61987]: DEBUG nova.scheduler.client.report [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 545.486107] env[61987]: DEBUG oslo_concurrency.lockutils [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Releasing lock "refresh_cache-31c175f9-20a5-4e38-ba90-972b847ab65e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.486107] env[61987]: DEBUG nova.compute.manager [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 545.486283] env[61987]: DEBUG nova.compute.manager [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 545.486377] env[61987]: DEBUG nova.network.neutron [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 545.534362] env[61987]: DEBUG nova.network.neutron [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 545.579086] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.599239] env[61987]: DEBUG nova.network.neutron [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.701409] env[61987]: ERROR nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5de8b27a-666f-4c53-9ef8-4b92e5c5a592, please check neutron logs for more information. [ 545.701409] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 545.701409] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 545.701409] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 545.701409] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 545.701409] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 545.701409] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 545.701409] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 545.701409] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.701409] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 545.701409] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.701409] env[61987]: ERROR nova.compute.manager raise self.value [ 545.701409] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 545.701409] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 545.701409] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.701409] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 545.701895] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.701895] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 545.701895] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5de8b27a-666f-4c53-9ef8-4b92e5c5a592, please check neutron logs for more information. [ 545.701895] env[61987]: ERROR nova.compute.manager [ 545.702026] env[61987]: Traceback (most recent call last): [ 545.702144] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 545.702144] env[61987]: listener.cb(fileno) [ 545.702144] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.702144] env[61987]: result = function(*args, **kwargs) [ 545.702144] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 545.702144] env[61987]: return func(*args, **kwargs) [ 545.702144] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 545.702144] env[61987]: raise e [ 545.702144] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 545.702144] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 545.702144] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 545.702144] env[61987]: created_port_ids = self._update_ports_for_instance( [ 545.702144] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 545.702144] env[61987]: with excutils.save_and_reraise_exception(): [ 545.702144] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.702144] env[61987]: self.force_reraise() [ 545.702144] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.702144] env[61987]: raise self.value [ 545.702144] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 545.702144] env[61987]: updated_port = self._update_port( [ 545.702144] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.702144] env[61987]: _ensure_no_port_binding_failure(port) [ 545.702144] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.702144] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 545.702144] env[61987]: nova.exception.PortBindingFailed: Binding failed for port 5de8b27a-666f-4c53-9ef8-4b92e5c5a592, please check neutron logs for more information. [ 545.702144] env[61987]: Removing descriptor: 17 [ 545.706516] env[61987]: ERROR nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5de8b27a-666f-4c53-9ef8-4b92e5c5a592, please check neutron logs for more information. [ 545.706516] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Traceback (most recent call last): [ 545.706516] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 545.706516] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] yield resources [ 545.706516] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 545.706516] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] self.driver.spawn(context, instance, image_meta, [ 545.706516] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 545.706516] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] self._vmops.spawn(context, instance, image_meta, injected_files, [ 545.706516] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 545.706516] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] vm_ref = self.build_virtual_machine(instance, [ 545.706516] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 545.707304] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] vif_infos = vmwarevif.get_vif_info(self._session, [ 545.707304] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 545.707304] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] for vif in network_info: [ 545.707304] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 545.707304] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] return self._sync_wrapper(fn, *args, **kwargs) [ 545.707304] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 545.707304] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] self.wait() [ 545.707304] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 545.707304] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] self[:] = self._gt.wait() [ 545.707304] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 545.707304] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] return self._exit_event.wait() [ 545.707304] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 545.707304] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] result = hub.switch() [ 545.707643] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 545.707643] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] return self.greenlet.switch() [ 545.707643] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.707643] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] result = function(*args, **kwargs) [ 545.707643] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 545.707643] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] return func(*args, **kwargs) [ 545.707643] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 545.707643] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] raise e [ 545.707643] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 545.707643] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] nwinfo = self.network_api.allocate_for_instance( [ 545.707643] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 545.707643] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] created_port_ids = self._update_ports_for_instance( [ 545.707643] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 545.707983] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] with excutils.save_and_reraise_exception(): [ 545.707983] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.707983] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] self.force_reraise() [ 545.707983] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.707983] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] raise self.value [ 545.707983] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 545.707983] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] updated_port = self._update_port( [ 545.707983] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.707983] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] _ensure_no_port_binding_failure(port) [ 545.707983] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.707983] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] raise exception.PortBindingFailed(port_id=port['id']) [ 545.707983] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] nova.exception.PortBindingFailed: Binding failed for port 5de8b27a-666f-4c53-9ef8-4b92e5c5a592, please check neutron logs for more information. [ 545.707983] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] [ 545.708326] env[61987]: INFO nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Terminating instance [ 545.992463] env[61987]: DEBUG oslo_concurrency.lockutils [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.910s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.993123] env[61987]: ERROR nova.compute.manager [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3f3ca432-e7a4-4a2f-9086-3983a2f72e29, please check neutron logs for more information. [ 545.993123] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Traceback (most recent call last): [ 545.993123] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 545.993123] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] self.driver.spawn(context, instance, image_meta, [ 545.993123] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 545.993123] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] self._vmops.spawn(context, instance, image_meta, injected_files, [ 545.993123] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 545.993123] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] vm_ref = self.build_virtual_machine(instance, [ 545.993123] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 545.993123] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] vif_infos = vmwarevif.get_vif_info(self._session, [ 545.993123] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 545.993562] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] for vif in network_info: [ 545.993562] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 545.993562] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] return self._sync_wrapper(fn, *args, **kwargs) [ 545.993562] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 545.993562] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] self.wait() [ 545.993562] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 545.993562] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] self[:] = self._gt.wait() [ 545.993562] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 545.993562] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] return self._exit_event.wait() [ 545.993562] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 545.993562] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] result = hub.switch() [ 545.993562] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 545.993562] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] return self.greenlet.switch() [ 545.994023] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.994023] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] result = function(*args, **kwargs) [ 545.994023] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 545.994023] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] return func(*args, **kwargs) [ 545.994023] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 545.994023] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] raise e [ 545.994023] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 545.994023] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] nwinfo = self.network_api.allocate_for_instance( [ 545.994023] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 545.994023] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] created_port_ids = self._update_ports_for_instance( [ 545.994023] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 545.994023] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] with excutils.save_and_reraise_exception(): [ 545.994023] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.994502] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] self.force_reraise() [ 545.994502] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.994502] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] raise self.value [ 545.994502] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 545.994502] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] updated_port = self._update_port( [ 545.994502] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.994502] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] _ensure_no_port_binding_failure(port) [ 545.994502] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.994502] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] raise exception.PortBindingFailed(port_id=port['id']) [ 545.994502] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] nova.exception.PortBindingFailed: Binding failed for port 3f3ca432-e7a4-4a2f-9086-3983a2f72e29, please check neutron logs for more information. [ 545.994502] env[61987]: ERROR nova.compute.manager [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] [ 545.995313] env[61987]: DEBUG nova.compute.utils [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Binding failed for port 3f3ca432-e7a4-4a2f-9086-3983a2f72e29, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 545.995890] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 13.999s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.995890] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.995890] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61987) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 545.996104] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.899s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.000211] env[61987]: DEBUG nova.compute.manager [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Build of instance 25f011ba-eb0b-4295-be8a-d223584ada63 was re-scheduled: Binding failed for port 3f3ca432-e7a4-4a2f-9086-3983a2f72e29, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 546.000365] env[61987]: DEBUG nova.compute.manager [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 546.000584] env[61987]: DEBUG oslo_concurrency.lockutils [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Acquiring lock "refresh_cache-25f011ba-eb0b-4295-be8a-d223584ada63" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.000728] env[61987]: DEBUG oslo_concurrency.lockutils [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Acquired lock "refresh_cache-25f011ba-eb0b-4295-be8a-d223584ada63" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.000880] env[61987]: DEBUG nova.network.neutron [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 546.002978] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-562e4ab6-506e-4fdc-9a3a-d49fb4f0a564 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.011446] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1215b61e-d8f9-4185-8070-36d34d126b6a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.030431] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1bb599e-e866-4eeb-b80f-f76b62e7f65d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.040426] env[61987]: DEBUG nova.network.neutron [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.044044] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-901a219e-a75d-41a3-9bcc-a99c60cec01c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.080398] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181577MB free_disk=180GB free_vcpus=48 pci_devices=None {{(pid=61987) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 546.080398] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.082395] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Releasing lock "refresh_cache-1a9030ce-3a35-4760-96f4-4288c1d2f9d1" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.082930] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 546.083069] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 546.083878] env[61987]: DEBUG oslo_concurrency.lockutils [req-7827b610-e0d2-4f78-8be3-6fc97638aefc req-cf652f89-d431-4fbd-a106-bfd03c2ce98d service nova] Acquired lock "refresh_cache-1a9030ce-3a35-4760-96f4-4288c1d2f9d1" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.083878] env[61987]: DEBUG nova.network.neutron [req-7827b610-e0d2-4f78-8be3-6fc97638aefc req-cf652f89-d431-4fbd-a106-bfd03c2ce98d service nova] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Refreshing network info cache for port 210424e1-1a59-4cf6-b39d-6ada30cc40e7 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 546.085461] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c2745bba-98ea-48cf-9c19-74dc749f09f9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.095129] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6771b73c-f5d7-4aae-b342-d2b7c6dc3dad {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.109775] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Releasing lock "refresh_cache-9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.109775] env[61987]: DEBUG nova.compute.manager [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 546.109775] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 546.109775] env[61987]: DEBUG oslo_concurrency.lockutils [req-53a20691-729d-43ae-bf0a-763b9cbf21d9 req-155f83e5-c508-4850-9b19-fa5c217a50d3 service nova] Acquired lock "refresh_cache-9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.109775] env[61987]: DEBUG nova.network.neutron [req-53a20691-729d-43ae-bf0a-763b9cbf21d9 req-155f83e5-c508-4850-9b19-fa5c217a50d3 service nova] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Refreshing network info cache for port 555557bc-6f57-400e-8dbb-216d588433f1 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 546.109958] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dee3f9f4-df36-4c50-bb81-383384e2a99d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.124619] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1a9030ce-3a35-4760-96f4-4288c1d2f9d1 could not be found. [ 546.124619] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 546.124619] env[61987]: INFO nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Took 0.04 seconds to destroy the instance on the hypervisor. [ 546.124856] env[61987]: DEBUG oslo.service.loopingcall [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 546.125503] env[61987]: DEBUG nova.compute.manager [-] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 546.125683] env[61987]: DEBUG nova.network.neutron [-] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 546.132152] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c76698bf-bad7-4642-a385-0e690e225257 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.153674] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3 could not be found. [ 546.153674] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 546.153858] env[61987]: INFO nova.compute.manager [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Took 0.05 seconds to destroy the instance on the hypervisor. [ 546.154579] env[61987]: DEBUG oslo.service.loopingcall [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 546.154579] env[61987]: DEBUG nova.compute.manager [-] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 546.154579] env[61987]: DEBUG nova.network.neutron [-] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 546.159798] env[61987]: DEBUG nova.network.neutron [-] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.209921] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquiring lock "refresh_cache-510803e9-89ce-4452-b245-ee1f124ef081" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.210552] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquired lock "refresh_cache-510803e9-89ce-4452-b245-ee1f124ef081" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.210552] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 546.320907] env[61987]: DEBUG nova.network.neutron [-] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.546337] env[61987]: INFO nova.compute.manager [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 31c175f9-20a5-4e38-ba90-972b847ab65e] Took 1.06 seconds to deallocate network for instance. [ 546.552622] env[61987]: DEBUG nova.network.neutron [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.560222] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Acquiring lock "660e5e4c-0804-429b-ab18-9305a433391f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.560222] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Lock "660e5e4c-0804-429b-ab18-9305a433391f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.662564] env[61987]: DEBUG nova.network.neutron [-] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.671946] env[61987]: DEBUG nova.network.neutron [req-7827b610-e0d2-4f78-8be3-6fc97638aefc req-cf652f89-d431-4fbd-a106-bfd03c2ce98d service nova] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.685954] env[61987]: DEBUG nova.network.neutron [req-53a20691-729d-43ae-bf0a-763b9cbf21d9 req-155f83e5-c508-4850-9b19-fa5c217a50d3 service nova] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.761497] env[61987]: DEBUG nova.network.neutron [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.776184] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.799585] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99c2deaf-30b1-402b-9dbf-c59d03fba9e4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.808248] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-609426b1-25da-4d56-b756-c133e78661ff {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.843033] env[61987]: DEBUG nova.network.neutron [-] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.845519] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ac26ce2-b8e0-4c3d-b050-f56e6add4567 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.854486] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc041368-159d-47ff-9614-e1e493836354 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.871795] env[61987]: DEBUG nova.compute.provider_tree [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 547.165091] env[61987]: INFO nova.compute.manager [-] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Took 1.04 seconds to deallocate network for instance. [ 547.168190] env[61987]: DEBUG nova.compute.claims [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 547.168436] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.200296] env[61987]: DEBUG nova.network.neutron [req-7827b610-e0d2-4f78-8be3-6fc97638aefc req-cf652f89-d431-4fbd-a106-bfd03c2ce98d service nova] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.231616] env[61987]: DEBUG nova.network.neutron [req-53a20691-729d-43ae-bf0a-763b9cbf21d9 req-155f83e5-c508-4850-9b19-fa5c217a50d3 service nova] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.264747] env[61987]: DEBUG oslo_concurrency.lockutils [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Releasing lock "refresh_cache-25f011ba-eb0b-4295-be8a-d223584ada63" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.264747] env[61987]: DEBUG nova.compute.manager [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 547.264747] env[61987]: DEBUG nova.compute.manager [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 547.265100] env[61987]: DEBUG nova.network.neutron [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 547.315379] env[61987]: DEBUG nova.network.neutron [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 547.318605] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.346859] env[61987]: INFO nova.compute.manager [-] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Took 1.19 seconds to deallocate network for instance. [ 547.348921] env[61987]: DEBUG nova.compute.claims [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 547.349175] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.377168] env[61987]: DEBUG nova.scheduler.client.report [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 547.597575] env[61987]: INFO nova.scheduler.client.report [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Deleted allocations for instance 31c175f9-20a5-4e38-ba90-972b847ab65e [ 547.703976] env[61987]: DEBUG oslo_concurrency.lockutils [req-7827b610-e0d2-4f78-8be3-6fc97638aefc req-cf652f89-d431-4fbd-a106-bfd03c2ce98d service nova] Releasing lock "refresh_cache-1a9030ce-3a35-4760-96f4-4288c1d2f9d1" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.740659] env[61987]: DEBUG oslo_concurrency.lockutils [req-53a20691-729d-43ae-bf0a-763b9cbf21d9 req-155f83e5-c508-4850-9b19-fa5c217a50d3 service nova] Releasing lock "refresh_cache-9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.818063] env[61987]: DEBUG nova.network.neutron [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.821593] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Releasing lock "refresh_cache-510803e9-89ce-4452-b245-ee1f124ef081" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.821991] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 547.823036] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 547.823036] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5d309aa9-3749-4ef7-be16-d1bec1f3bc05 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.832546] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c87ec36-c7fd-41aa-a75b-a1d0bdec31ec {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.859101] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 510803e9-89ce-4452-b245-ee1f124ef081 could not be found. [ 547.859324] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 547.859498] env[61987]: INFO nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Took 0.04 seconds to destroy the instance on the hypervisor. [ 547.859749] env[61987]: DEBUG oslo.service.loopingcall [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 547.859968] env[61987]: DEBUG nova.compute.manager [-] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 547.860068] env[61987]: DEBUG nova.network.neutron [-] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 547.883922] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.888s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.884659] env[61987]: ERROR nova.compute.manager [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cd2e6086-23ab-4064-a665-7e4b92b07e6f, please check neutron logs for more information. [ 547.884659] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Traceback (most recent call last): [ 547.884659] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 547.884659] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] self.driver.spawn(context, instance, image_meta, [ 547.884659] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 547.884659] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] self._vmops.spawn(context, instance, image_meta, injected_files, [ 547.884659] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 547.884659] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] vm_ref = self.build_virtual_machine(instance, [ 547.884659] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 547.884659] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] vif_infos = vmwarevif.get_vif_info(self._session, [ 547.884659] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 547.884996] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] for vif in network_info: [ 547.884996] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 547.884996] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] return self._sync_wrapper(fn, *args, **kwargs) [ 547.884996] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 547.884996] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] self.wait() [ 547.884996] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 547.884996] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] self[:] = self._gt.wait() [ 547.884996] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 547.884996] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] return self._exit_event.wait() [ 547.884996] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 547.884996] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] result = hub.switch() [ 547.884996] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 547.884996] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] return self.greenlet.switch() [ 547.885317] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.885317] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] result = function(*args, **kwargs) [ 547.885317] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 547.885317] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] return func(*args, **kwargs) [ 547.885317] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 547.885317] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] raise e [ 547.885317] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 547.885317] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] nwinfo = self.network_api.allocate_for_instance( [ 547.885317] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 547.885317] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] created_port_ids = self._update_ports_for_instance( [ 547.885317] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 547.885317] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] with excutils.save_and_reraise_exception(): [ 547.885317] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.885624] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] self.force_reraise() [ 547.885624] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.885624] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] raise self.value [ 547.885624] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 547.885624] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] updated_port = self._update_port( [ 547.885624] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.885624] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] _ensure_no_port_binding_failure(port) [ 547.885624] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.885624] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] raise exception.PortBindingFailed(port_id=port['id']) [ 547.885624] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] nova.exception.PortBindingFailed: Binding failed for port cd2e6086-23ab-4064-a665-7e4b92b07e6f, please check neutron logs for more information. [ 547.885624] env[61987]: ERROR nova.compute.manager [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] [ 547.885893] env[61987]: DEBUG nova.compute.utils [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Binding failed for port cd2e6086-23ab-4064-a665-7e4b92b07e6f, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 547.888787] env[61987]: DEBUG nova.compute.manager [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Build of instance 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac was re-scheduled: Binding failed for port cd2e6086-23ab-4064-a665-7e4b92b07e6f, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 547.889718] env[61987]: DEBUG nova.compute.manager [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 547.890062] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Acquiring lock "refresh_cache-7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.890265] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Acquired lock "refresh_cache-7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.890465] env[61987]: DEBUG nova.network.neutron [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 547.891500] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.667s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.901957] env[61987]: DEBUG nova.network.neutron [-] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 548.036382] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Acquiring lock "c7bfab3b-bfb1-416f-a302-9956c4fb0748" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.037715] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Lock "c7bfab3b-bfb1-416f-a302-9956c4fb0748" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.089201] env[61987]: ERROR nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 30ce6103-be54-438d-930c-51e75dbcf857, please check neutron logs for more information. [ 548.089201] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 548.089201] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 548.089201] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 548.089201] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 548.089201] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 548.089201] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 548.089201] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 548.089201] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.089201] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 548.089201] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.089201] env[61987]: ERROR nova.compute.manager raise self.value [ 548.089201] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 548.089201] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 548.089201] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.089201] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 548.089726] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.089726] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 548.089726] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 30ce6103-be54-438d-930c-51e75dbcf857, please check neutron logs for more information. [ 548.089726] env[61987]: ERROR nova.compute.manager [ 548.089726] env[61987]: Traceback (most recent call last): [ 548.089726] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 548.089726] env[61987]: listener.cb(fileno) [ 548.089726] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.089726] env[61987]: result = function(*args, **kwargs) [ 548.089726] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 548.089726] env[61987]: return func(*args, **kwargs) [ 548.089726] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 548.089726] env[61987]: raise e [ 548.089726] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 548.089726] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 548.089726] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 548.089726] env[61987]: created_port_ids = self._update_ports_for_instance( [ 548.089726] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 548.089726] env[61987]: with excutils.save_and_reraise_exception(): [ 548.089726] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.089726] env[61987]: self.force_reraise() [ 548.089726] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.089726] env[61987]: raise self.value [ 548.089726] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 548.089726] env[61987]: updated_port = self._update_port( [ 548.089726] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.089726] env[61987]: _ensure_no_port_binding_failure(port) [ 548.089726] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.089726] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 548.090567] env[61987]: nova.exception.PortBindingFailed: Binding failed for port 30ce6103-be54-438d-930c-51e75dbcf857, please check neutron logs for more information. [ 548.090567] env[61987]: Removing descriptor: 18 [ 548.090567] env[61987]: ERROR nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 30ce6103-be54-438d-930c-51e75dbcf857, please check neutron logs for more information. [ 548.090567] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Traceback (most recent call last): [ 548.090567] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 548.090567] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] yield resources [ 548.090567] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 548.090567] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] self.driver.spawn(context, instance, image_meta, [ 548.090567] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 548.090567] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 548.090567] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 548.090567] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] vm_ref = self.build_virtual_machine(instance, [ 548.090936] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 548.090936] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] vif_infos = vmwarevif.get_vif_info(self._session, [ 548.090936] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 548.090936] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] for vif in network_info: [ 548.090936] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 548.090936] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] return self._sync_wrapper(fn, *args, **kwargs) [ 548.090936] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 548.090936] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] self.wait() [ 548.090936] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 548.090936] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] self[:] = self._gt.wait() [ 548.090936] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 548.090936] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] return self._exit_event.wait() [ 548.090936] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 548.091309] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] result = hub.switch() [ 548.091309] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 548.091309] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] return self.greenlet.switch() [ 548.091309] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.091309] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] result = function(*args, **kwargs) [ 548.091309] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 548.091309] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] return func(*args, **kwargs) [ 548.091309] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 548.091309] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] raise e [ 548.091309] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 548.091309] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] nwinfo = self.network_api.allocate_for_instance( [ 548.091309] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 548.091309] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] created_port_ids = self._update_ports_for_instance( [ 548.091644] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 548.091644] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] with excutils.save_and_reraise_exception(): [ 548.091644] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.091644] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] self.force_reraise() [ 548.091644] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.091644] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] raise self.value [ 548.091644] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 548.091644] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] updated_port = self._update_port( [ 548.091644] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.091644] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] _ensure_no_port_binding_failure(port) [ 548.091644] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.091644] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] raise exception.PortBindingFailed(port_id=port['id']) [ 548.091993] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] nova.exception.PortBindingFailed: Binding failed for port 30ce6103-be54-438d-930c-51e75dbcf857, please check neutron logs for more information. [ 548.091993] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] [ 548.091993] env[61987]: INFO nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Terminating instance [ 548.109888] env[61987]: DEBUG oslo_concurrency.lockutils [None req-95019981-c991-4482-bcc7-80e8203cc97f tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "31c175f9-20a5-4e38-ba90-972b847ab65e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.747s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.323935] env[61987]: INFO nova.compute.manager [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] [instance: 25f011ba-eb0b-4295-be8a-d223584ada63] Took 1.06 seconds to deallocate network for instance. [ 548.406791] env[61987]: DEBUG nova.network.neutron [-] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.439298] env[61987]: DEBUG nova.network.neutron [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 548.552086] env[61987]: DEBUG nova.network.neutron [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.578210] env[61987]: DEBUG nova.compute.manager [req-90161d32-99ba-4419-8dba-c4da507cf391 req-1cbaa5ea-c3a5-45c9-90d1-cc8cb8501315 service nova] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Received event network-vif-deleted-210424e1-1a59-4cf6-b39d-6ada30cc40e7 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 548.596540] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquiring lock "refresh_cache-c9ce2911-8fc4-4aa8-883e-487ee06670ba" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.597676] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquired lock "refresh_cache-c9ce2911-8fc4-4aa8-883e-487ee06670ba" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.597974] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 548.605337] env[61987]: DEBUG nova.compute.manager [req-1de64f46-3aec-42a3-93d1-ef89e5e20392 req-9d8560cf-c28f-4d47-b86b-258c0439714f service nova] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Received event network-vif-deleted-555557bc-6f57-400e-8dbb-216d588433f1 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 548.605426] env[61987]: DEBUG nova.compute.manager [req-1de64f46-3aec-42a3-93d1-ef89e5e20392 req-9d8560cf-c28f-4d47-b86b-258c0439714f service nova] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Received event network-changed-5de8b27a-666f-4c53-9ef8-4b92e5c5a592 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 548.605535] env[61987]: DEBUG nova.compute.manager [req-1de64f46-3aec-42a3-93d1-ef89e5e20392 req-9d8560cf-c28f-4d47-b86b-258c0439714f service nova] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Refreshing instance network info cache due to event network-changed-5de8b27a-666f-4c53-9ef8-4b92e5c5a592. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 548.605731] env[61987]: DEBUG oslo_concurrency.lockutils [req-1de64f46-3aec-42a3-93d1-ef89e5e20392 req-9d8560cf-c28f-4d47-b86b-258c0439714f service nova] Acquiring lock "refresh_cache-510803e9-89ce-4452-b245-ee1f124ef081" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.605862] env[61987]: DEBUG oslo_concurrency.lockutils [req-1de64f46-3aec-42a3-93d1-ef89e5e20392 req-9d8560cf-c28f-4d47-b86b-258c0439714f service nova] Acquired lock "refresh_cache-510803e9-89ce-4452-b245-ee1f124ef081" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.607215] env[61987]: DEBUG nova.network.neutron [req-1de64f46-3aec-42a3-93d1-ef89e5e20392 req-9d8560cf-c28f-4d47-b86b-258c0439714f service nova] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Refreshing network info cache for port 5de8b27a-666f-4c53-9ef8-4b92e5c5a592 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 548.614622] env[61987]: DEBUG nova.compute.manager [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 548.696950] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b157da1-ab70-40af-a13d-2c6b2e9db0d0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.705629] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4827b6d-bf55-4efd-b478-53c0a15db9a5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.737142] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57255a89-6a04-4eda-aa18-b7d9eb6f7c6b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.744571] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a665696-23d0-494f-8b7a-d73afc6e6c3e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.758803] env[61987]: DEBUG nova.compute.provider_tree [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 548.909365] env[61987]: INFO nova.compute.manager [-] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Took 1.05 seconds to deallocate network for instance. [ 548.912984] env[61987]: DEBUG nova.compute.claims [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 548.913176] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.056760] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Releasing lock "refresh_cache-7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.057015] env[61987]: DEBUG nova.compute.manager [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 549.057196] env[61987]: DEBUG nova.compute.manager [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 549.057391] env[61987]: DEBUG nova.network.neutron [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 549.090829] env[61987]: DEBUG nova.network.neutron [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 549.143961] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 549.146279] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.175589] env[61987]: DEBUG nova.network.neutron [req-1de64f46-3aec-42a3-93d1-ef89e5e20392 req-9d8560cf-c28f-4d47-b86b-258c0439714f service nova] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 549.261713] env[61987]: DEBUG nova.scheduler.client.report [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 549.378679] env[61987]: INFO nova.scheduler.client.report [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Deleted allocations for instance 25f011ba-eb0b-4295-be8a-d223584ada63 [ 549.594818] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.599496] env[61987]: DEBUG nova.network.neutron [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.626048] env[61987]: DEBUG nova.network.neutron [req-1de64f46-3aec-42a3-93d1-ef89e5e20392 req-9d8560cf-c28f-4d47-b86b-258c0439714f service nova] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.767375] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.876s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.769186] env[61987]: ERROR nova.compute.manager [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f80fb674-a7f0-4ea0-85b8-a7b404d94b8c, please check neutron logs for more information. [ 549.769186] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Traceback (most recent call last): [ 549.769186] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 549.769186] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] self.driver.spawn(context, instance, image_meta, [ 549.769186] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 549.769186] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 549.769186] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 549.769186] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] vm_ref = self.build_virtual_machine(instance, [ 549.769186] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 549.769186] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] vif_infos = vmwarevif.get_vif_info(self._session, [ 549.769186] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 549.769522] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] for vif in network_info: [ 549.769522] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 549.769522] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] return self._sync_wrapper(fn, *args, **kwargs) [ 549.769522] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 549.769522] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] self.wait() [ 549.769522] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 549.769522] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] self[:] = self._gt.wait() [ 549.769522] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 549.769522] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] return self._exit_event.wait() [ 549.769522] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 549.769522] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] result = hub.switch() [ 549.769522] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 549.769522] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] return self.greenlet.switch() [ 549.769838] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.769838] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] result = function(*args, **kwargs) [ 549.769838] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 549.769838] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] return func(*args, **kwargs) [ 549.769838] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 549.769838] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] raise e [ 549.769838] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 549.769838] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] nwinfo = self.network_api.allocate_for_instance( [ 549.769838] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 549.769838] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] created_port_ids = self._update_ports_for_instance( [ 549.769838] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 549.769838] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] with excutils.save_and_reraise_exception(): [ 549.769838] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.770172] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] self.force_reraise() [ 549.770172] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.770172] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] raise self.value [ 549.770172] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 549.770172] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] updated_port = self._update_port( [ 549.770172] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.770172] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] _ensure_no_port_binding_failure(port) [ 549.770172] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.770172] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] raise exception.PortBindingFailed(port_id=port['id']) [ 549.770172] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] nova.exception.PortBindingFailed: Binding failed for port f80fb674-a7f0-4ea0-85b8-a7b404d94b8c, please check neutron logs for more information. [ 549.770172] env[61987]: ERROR nova.compute.manager [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] [ 549.770687] env[61987]: DEBUG nova.compute.utils [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Binding failed for port f80fb674-a7f0-4ea0-85b8-a7b404d94b8c, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 549.771976] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.686s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.774957] env[61987]: DEBUG nova.compute.manager [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Build of instance ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6 was re-scheduled: Binding failed for port f80fb674-a7f0-4ea0-85b8-a7b404d94b8c, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 549.775439] env[61987]: DEBUG nova.compute.manager [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 549.775993] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Acquiring lock "refresh_cache-ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.775993] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Acquired lock "refresh_cache-ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.775993] env[61987]: DEBUG nova.network.neutron [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 549.891421] env[61987]: DEBUG oslo_concurrency.lockutils [None req-973e28a0-71b7-4942-a793-f64ef84d0fa2 tempest-ServerDiagnosticsNegativeTest-493040276 tempest-ServerDiagnosticsNegativeTest-493040276-project-member] Lock "25f011ba-eb0b-4295-be8a-d223584ada63" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.714s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.098625] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Releasing lock "refresh_cache-c9ce2911-8fc4-4aa8-883e-487ee06670ba" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.099393] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 550.099612] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 550.100022] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f1119e6e-6c7f-4c54-8db8-44f7329914d3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.106304] env[61987]: INFO nova.compute.manager [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac] Took 1.05 seconds to deallocate network for instance. [ 550.118823] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb75daa-d74b-404d-b22d-436495ce9a15 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.129558] env[61987]: DEBUG oslo_concurrency.lockutils [req-1de64f46-3aec-42a3-93d1-ef89e5e20392 req-9d8560cf-c28f-4d47-b86b-258c0439714f service nova] Releasing lock "refresh_cache-510803e9-89ce-4452-b245-ee1f124ef081" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.129558] env[61987]: DEBUG nova.compute.manager [req-1de64f46-3aec-42a3-93d1-ef89e5e20392 req-9d8560cf-c28f-4d47-b86b-258c0439714f service nova] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Received event network-vif-deleted-5de8b27a-666f-4c53-9ef8-4b92e5c5a592 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 550.143586] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c9ce2911-8fc4-4aa8-883e-487ee06670ba could not be found. [ 550.143826] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 550.144037] env[61987]: INFO nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Took 0.04 seconds to destroy the instance on the hypervisor. [ 550.144369] env[61987]: DEBUG oslo.service.loopingcall [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 550.144717] env[61987]: DEBUG nova.compute.manager [-] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 550.144717] env[61987]: DEBUG nova.network.neutron [-] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 550.179591] env[61987]: DEBUG nova.network.neutron [-] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 550.314096] env[61987]: DEBUG nova.network.neutron [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 550.395797] env[61987]: DEBUG nova.compute.manager [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 550.590685] env[61987]: DEBUG nova.network.neutron [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.631914] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7093270b-12bf-4117-9c8a-c5b695d6ef76 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.645559] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c6cd0a4-2d61-46f4-95d0-43e246380780 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.684618] env[61987]: DEBUG nova.network.neutron [-] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.686416] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71ceef40-64f1-4654-b51d-23211ed0a216 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.695777] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d1de1b6-d329-4a16-9f28-650efa118e17 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.715197] env[61987]: DEBUG nova.compute.provider_tree [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 550.923040] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.097735] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Releasing lock "refresh_cache-ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.098018] env[61987]: DEBUG nova.compute.manager [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 551.098255] env[61987]: DEBUG nova.compute.manager [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 551.098642] env[61987]: DEBUG nova.network.neutron [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 551.137912] env[61987]: DEBUG nova.network.neutron [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 551.147696] env[61987]: INFO nova.scheduler.client.report [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Deleted allocations for instance 7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac [ 551.191360] env[61987]: INFO nova.compute.manager [-] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Took 1.05 seconds to deallocate network for instance. [ 551.194904] env[61987]: DEBUG nova.compute.claims [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 551.195343] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.219415] env[61987]: DEBUG nova.scheduler.client.report [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 551.642612] env[61987]: DEBUG nova.network.neutron [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.661366] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa2ca8ab-468b-412f-909f-cdd965c454eb tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Lock "7d9aa694-f9b3-4d2b-9d3c-de59b8b5a6ac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.756s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.689413] env[61987]: DEBUG nova.compute.manager [req-20a953ee-29fb-477d-99a7-330026c1bd35 req-bdce0007-867e-4efa-a0bb-92b2649ff7cc service nova] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Received event network-changed-30ce6103-be54-438d-930c-51e75dbcf857 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 551.690458] env[61987]: DEBUG nova.compute.manager [req-20a953ee-29fb-477d-99a7-330026c1bd35 req-bdce0007-867e-4efa-a0bb-92b2649ff7cc service nova] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Refreshing instance network info cache due to event network-changed-30ce6103-be54-438d-930c-51e75dbcf857. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 551.690458] env[61987]: DEBUG oslo_concurrency.lockutils [req-20a953ee-29fb-477d-99a7-330026c1bd35 req-bdce0007-867e-4efa-a0bb-92b2649ff7cc service nova] Acquiring lock "refresh_cache-c9ce2911-8fc4-4aa8-883e-487ee06670ba" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.690458] env[61987]: DEBUG oslo_concurrency.lockutils [req-20a953ee-29fb-477d-99a7-330026c1bd35 req-bdce0007-867e-4efa-a0bb-92b2649ff7cc service nova] Acquired lock "refresh_cache-c9ce2911-8fc4-4aa8-883e-487ee06670ba" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.690458] env[61987]: DEBUG nova.network.neutron [req-20a953ee-29fb-477d-99a7-330026c1bd35 req-bdce0007-867e-4efa-a0bb-92b2649ff7cc service nova] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Refreshing network info cache for port 30ce6103-be54-438d-930c-51e75dbcf857 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 551.726573] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.955s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.727232] env[61987]: ERROR nova.compute.manager [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 516b0723-ba96-4f21-9f78-8765da5d19a8, please check neutron logs for more information. [ 551.727232] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Traceback (most recent call last): [ 551.727232] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 551.727232] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] self.driver.spawn(context, instance, image_meta, [ 551.727232] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 551.727232] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 551.727232] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 551.727232] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] vm_ref = self.build_virtual_machine(instance, [ 551.727232] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 551.727232] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] vif_infos = vmwarevif.get_vif_info(self._session, [ 551.727232] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 551.727594] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] for vif in network_info: [ 551.727594] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 551.727594] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] return self._sync_wrapper(fn, *args, **kwargs) [ 551.727594] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 551.727594] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] self.wait() [ 551.727594] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 551.727594] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] self[:] = self._gt.wait() [ 551.727594] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 551.727594] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] return self._exit_event.wait() [ 551.727594] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 551.727594] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] result = hub.switch() [ 551.727594] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 551.727594] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] return self.greenlet.switch() [ 551.727939] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.727939] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] result = function(*args, **kwargs) [ 551.727939] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.727939] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] return func(*args, **kwargs) [ 551.727939] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 551.727939] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] raise e [ 551.727939] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 551.727939] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] nwinfo = self.network_api.allocate_for_instance( [ 551.727939] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 551.727939] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] created_port_ids = self._update_ports_for_instance( [ 551.727939] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 551.727939] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] with excutils.save_and_reraise_exception(): [ 551.727939] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.729161] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] self.force_reraise() [ 551.729161] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.729161] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] raise self.value [ 551.729161] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 551.729161] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] updated_port = self._update_port( [ 551.729161] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.729161] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] _ensure_no_port_binding_failure(port) [ 551.729161] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.729161] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] raise exception.PortBindingFailed(port_id=port['id']) [ 551.729161] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] nova.exception.PortBindingFailed: Binding failed for port 516b0723-ba96-4f21-9f78-8765da5d19a8, please check neutron logs for more information. [ 551.729161] env[61987]: ERROR nova.compute.manager [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] [ 551.729456] env[61987]: DEBUG nova.compute.utils [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Binding failed for port 516b0723-ba96-4f21-9f78-8765da5d19a8, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 551.730370] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 7.751s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.737784] env[61987]: DEBUG nova.compute.manager [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Build of instance 71316e2e-cda6-461c-8004-ff80fc2da90c was re-scheduled: Binding failed for port 516b0723-ba96-4f21-9f78-8765da5d19a8, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 551.738113] env[61987]: DEBUG nova.compute.manager [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 551.739133] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Acquiring lock "refresh_cache-71316e2e-cda6-461c-8004-ff80fc2da90c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.739133] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Acquired lock "refresh_cache-71316e2e-cda6-461c-8004-ff80fc2da90c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.739133] env[61987]: DEBUG nova.network.neutron [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 552.144940] env[61987]: INFO nova.compute.manager [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] [instance: ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6] Took 1.05 seconds to deallocate network for instance. [ 552.167823] env[61987]: DEBUG nova.compute.manager [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 552.217781] env[61987]: DEBUG nova.network.neutron [req-20a953ee-29fb-477d-99a7-330026c1bd35 req-bdce0007-867e-4efa-a0bb-92b2649ff7cc service nova] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 552.270681] env[61987]: DEBUG nova.network.neutron [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 552.399446] env[61987]: DEBUG nova.network.neutron [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.410904] env[61987]: DEBUG nova.network.neutron [req-20a953ee-29fb-477d-99a7-330026c1bd35 req-bdce0007-867e-4efa-a0bb-92b2649ff7cc service nova] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.500365] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4694e5d-200d-4ba9-bf77-d76ebbedb532 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.508305] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2c0383-da67-40b4-9df8-95d8bb58b5aa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.538743] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a06c695-2a3e-49b7-a6e9-c778308463ea {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.546229] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-183fd2eb-d42f-423e-9272-d02bd3ce1760 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.559543] env[61987]: DEBUG nova.compute.provider_tree [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 552.692917] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.902422] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Releasing lock "refresh_cache-71316e2e-cda6-461c-8004-ff80fc2da90c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.902674] env[61987]: DEBUG nova.compute.manager [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 552.902922] env[61987]: DEBUG nova.compute.manager [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 552.903219] env[61987]: DEBUG nova.network.neutron [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 552.916222] env[61987]: DEBUG oslo_concurrency.lockutils [req-20a953ee-29fb-477d-99a7-330026c1bd35 req-bdce0007-867e-4efa-a0bb-92b2649ff7cc service nova] Releasing lock "refresh_cache-c9ce2911-8fc4-4aa8-883e-487ee06670ba" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.916222] env[61987]: DEBUG nova.compute.manager [req-20a953ee-29fb-477d-99a7-330026c1bd35 req-bdce0007-867e-4efa-a0bb-92b2649ff7cc service nova] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Received event network-vif-deleted-30ce6103-be54-438d-930c-51e75dbcf857 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 552.920409] env[61987]: DEBUG nova.network.neutron [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 552.951014] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "1915d6b8-2798-4e35-9a35-4191659c2cb6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.951379] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "1915d6b8-2798-4e35-9a35-4191659c2cb6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.068610] env[61987]: DEBUG nova.scheduler.client.report [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 553.186041] env[61987]: INFO nova.scheduler.client.report [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Deleted allocations for instance ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6 [ 553.200828] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Acquiring lock "e6c56d86-e428-49fb-9944-293ad5e5948c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.201280] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Lock "e6c56d86-e428-49fb-9944-293ad5e5948c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.423532] env[61987]: DEBUG nova.network.neutron [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.570714] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.841s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.572223] env[61987]: ERROR nova.compute.manager [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 76fe4435-5b40-4bf3-82ac-e4268f7a3ddd, please check neutron logs for more information. [ 553.572223] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Traceback (most recent call last): [ 553.572223] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 553.572223] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] self.driver.spawn(context, instance, image_meta, [ 553.572223] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 553.572223] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 553.572223] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 553.572223] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] vm_ref = self.build_virtual_machine(instance, [ 553.572223] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 553.572223] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] vif_infos = vmwarevif.get_vif_info(self._session, [ 553.572223] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 553.572983] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] for vif in network_info: [ 553.572983] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 553.572983] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] return self._sync_wrapper(fn, *args, **kwargs) [ 553.572983] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 553.572983] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] self.wait() [ 553.572983] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 553.572983] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] self[:] = self._gt.wait() [ 553.572983] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 553.572983] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] return self._exit_event.wait() [ 553.572983] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 553.572983] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] result = hub.switch() [ 553.572983] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 553.572983] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] return self.greenlet.switch() [ 553.573369] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.573369] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] result = function(*args, **kwargs) [ 553.573369] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 553.573369] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] return func(*args, **kwargs) [ 553.573369] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 553.573369] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] raise e [ 553.573369] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 553.573369] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] nwinfo = self.network_api.allocate_for_instance( [ 553.573369] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 553.573369] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] created_port_ids = self._update_ports_for_instance( [ 553.573369] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 553.573369] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] with excutils.save_and_reraise_exception(): [ 553.573369] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.573837] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] self.force_reraise() [ 553.573837] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.573837] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] raise self.value [ 553.573837] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 553.573837] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] updated_port = self._update_port( [ 553.573837] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.573837] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] _ensure_no_port_binding_failure(port) [ 553.573837] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.573837] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] raise exception.PortBindingFailed(port_id=port['id']) [ 553.573837] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] nova.exception.PortBindingFailed: Binding failed for port 76fe4435-5b40-4bf3-82ac-e4268f7a3ddd, please check neutron logs for more information. [ 553.573837] env[61987]: ERROR nova.compute.manager [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] [ 553.574682] env[61987]: DEBUG nova.compute.utils [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Binding failed for port 76fe4435-5b40-4bf3-82ac-e4268f7a3ddd, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 553.575558] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 7.496s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.577890] env[61987]: DEBUG nova.compute.manager [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Build of instance a0a004c3-6254-4f47-a080-2b9cc85a2a1d was re-scheduled: Binding failed for port 76fe4435-5b40-4bf3-82ac-e4268f7a3ddd, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 553.579029] env[61987]: DEBUG nova.compute.manager [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 553.579029] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Acquiring lock "refresh_cache-a0a004c3-6254-4f47-a080-2b9cc85a2a1d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.579029] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Acquired lock "refresh_cache-a0a004c3-6254-4f47-a080-2b9cc85a2a1d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.579029] env[61987]: DEBUG nova.network.neutron [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 553.703919] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce208087-020a-4d66-9d65-98cad5bb72f2 tempest-ServerMetadataNegativeTestJSON-479386389 tempest-ServerMetadataNegativeTestJSON-479386389-project-member] Lock "ed4e609d-e6ef-42d4-b5d5-89bdf99e36c6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.632s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.928470] env[61987]: INFO nova.compute.manager [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] [instance: 71316e2e-cda6-461c-8004-ff80fc2da90c] Took 1.03 seconds to deallocate network for instance. [ 554.009643] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "40fefd68-3d6f-4451-997f-e8fa49847dea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.009906] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "40fefd68-3d6f-4451-997f-e8fa49847dea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.013876] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Acquiring lock "c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.014114] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Lock "c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.119239] env[61987]: DEBUG nova.network.neutron [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 554.213207] env[61987]: DEBUG nova.compute.manager [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 554.574496] env[61987]: DEBUG nova.network.neutron [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.745075] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.971993] env[61987]: INFO nova.scheduler.client.report [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Deleted allocations for instance 71316e2e-cda6-461c-8004-ff80fc2da90c [ 555.077211] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Releasing lock "refresh_cache-a0a004c3-6254-4f47-a080-2b9cc85a2a1d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.077702] env[61987]: DEBUG nova.compute.manager [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 555.077918] env[61987]: DEBUG nova.compute.manager [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 555.081023] env[61987]: DEBUG nova.network.neutron [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 555.108192] env[61987]: DEBUG nova.network.neutron [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 555.122873] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 71316e2e-cda6-461c-8004-ff80fc2da90c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 555.485389] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f51a63b9-10c6-438d-b086-64c7c2b98789 tempest-ServersAdminNegativeTestJSON-1236920184 tempest-ServersAdminNegativeTestJSON-1236920184-project-member] Lock "71316e2e-cda6-461c-8004-ff80fc2da90c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.596s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.610642] env[61987]: DEBUG nova.network.neutron [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.628239] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance a0a004c3-6254-4f47-a080-2b9cc85a2a1d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 555.628914] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 555.630930] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 510803e9-89ce-4452-b245-ee1f124ef081 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 555.630930] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 1a9030ce-3a35-4760-96f4-4288c1d2f9d1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 555.630930] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance c9ce2911-8fc4-4aa8-883e-487ee06670ba actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 555.990100] env[61987]: DEBUG nova.compute.manager [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 556.114118] env[61987]: INFO nova.compute.manager [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] [instance: a0a004c3-6254-4f47-a080-2b9cc85a2a1d] Took 1.04 seconds to deallocate network for instance. [ 556.133494] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 556.516696] env[61987]: DEBUG oslo_concurrency.lockutils [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.640028] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 3182ca24-c989-4c29-8e47-d913d0607240 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 557.120841] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Acquiring lock "967bc997-543f-4b35-968b-c530b82d9db0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.121584] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Lock "967bc997-543f-4b35-968b-c530b82d9db0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.144398] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 7a94f776-579d-41e8-94aa-6095b3fd69c6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 557.154498] env[61987]: INFO nova.scheduler.client.report [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Deleted allocations for instance a0a004c3-6254-4f47-a080-2b9cc85a2a1d [ 557.650591] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 938dab0c-9b30-404b-bca2-a1280a252584 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 557.662654] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0b79b0e6-a682-40c2-b497-b9943f22462e tempest-ImagesNegativeTestJSON-1392753326 tempest-ImagesNegativeTestJSON-1392753326-project-member] Lock "a0a004c3-6254-4f47-a080-2b9cc85a2a1d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.543s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.137175] env[61987]: DEBUG oslo_concurrency.lockutils [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Acquiring lock "ab2edc70-b4c6-4db3-b0d6-53148764efdd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.137460] env[61987]: DEBUG oslo_concurrency.lockutils [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Lock "ab2edc70-b4c6-4db3-b0d6-53148764efdd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.155923] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 6bf0d1f1-d079-4956-bc9a-86ba66885f3a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 558.169564] env[61987]: DEBUG nova.compute.manager [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 558.662382] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 7961bd0d-de24-4e13-94a3-79c48c593813 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 558.700337] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.165261] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance bac5a3bc-5a58-4c1d-abb8-1aaf6151358a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 559.669103] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 660e5e4c-0804-429b-ab18-9305a433391f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 559.825410] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Acquiring lock "388b62aa-6647-4aca-acb2-8d7d7d424c8c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.825663] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Lock "388b62aa-6647-4aca-acb2-8d7d7d424c8c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.175605] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance c7bfab3b-bfb1-416f-a302-9956c4fb0748 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 560.681862] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 1915d6b8-2798-4e35-9a35-4191659c2cb6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 561.186843] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance e6c56d86-e428-49fb-9944-293ad5e5948c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 561.689833] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 562.192931] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 40fefd68-3d6f-4451-997f-e8fa49847dea has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 562.193839] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 562.194033] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 562.520104] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-077f937e-1592-4f24-bd32-9526bbb2d5c6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.528781] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3516dc9-0b22-4c1e-8fe8-b1e6f87bb34b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.564559] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a62c53f9-e207-4382-95e4-77e91d0b1dac {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.573448] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de709161-5480-48f3-a70b-79e2135f460a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.591200] env[61987]: DEBUG nova.compute.provider_tree [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 563.095050] env[61987]: DEBUG nova.scheduler.client.report [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 563.606730] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61987) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 563.606730] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.029s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.606730] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.436s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.421377] env[61987]: DEBUG oslo_concurrency.lockutils [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Acquiring lock "96bbc732-cb58-4de2-80d3-c527228adaea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.421377] env[61987]: DEBUG oslo_concurrency.lockutils [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Lock "96bbc732-cb58-4de2-80d3-c527228adaea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.517729] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30429d0a-c81c-4689-9911-466ca838256a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.526759] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e1039fc-f745-4179-8f27-d4192ab5acc3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.561589] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c80a4fd6-3d35-433c-ab4d-09ba7b2f8639 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.569427] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc522049-b630-44d5-8f11-ebdce45fa830 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.585090] env[61987]: DEBUG nova.compute.provider_tree [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 565.088447] env[61987]: DEBUG nova.scheduler.client.report [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 565.544252] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "699c853a-a88f-4890-b712-00f00a6c9838" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.546645] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "699c853a-a88f-4890-b712-00f00a6c9838" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.573263] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "65ddff8e-971c-4485-b658-4611aea1176c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.573495] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "65ddff8e-971c-4485-b658-4611aea1176c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.597285] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.989s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.597285] env[61987]: ERROR nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 210424e1-1a59-4cf6-b39d-6ada30cc40e7, please check neutron logs for more information. [ 565.597285] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Traceback (most recent call last): [ 565.597285] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 565.597285] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] self.driver.spawn(context, instance, image_meta, [ 565.597285] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 565.597285] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 565.597285] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 565.597285] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] vm_ref = self.build_virtual_machine(instance, [ 565.597569] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 565.597569] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] vif_infos = vmwarevif.get_vif_info(self._session, [ 565.597569] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 565.597569] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] for vif in network_info: [ 565.597569] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 565.597569] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] return self._sync_wrapper(fn, *args, **kwargs) [ 565.597569] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 565.597569] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] self.wait() [ 565.597569] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 565.597569] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] self[:] = self._gt.wait() [ 565.597569] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 565.597569] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] return self._exit_event.wait() [ 565.597569] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 565.597930] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] result = hub.switch() [ 565.597930] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 565.597930] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] return self.greenlet.switch() [ 565.597930] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.597930] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] result = function(*args, **kwargs) [ 565.597930] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 565.597930] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] return func(*args, **kwargs) [ 565.597930] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 565.597930] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] raise e [ 565.597930] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 565.597930] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] nwinfo = self.network_api.allocate_for_instance( [ 565.597930] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 565.597930] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] created_port_ids = self._update_ports_for_instance( [ 565.598294] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 565.598294] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] with excutils.save_and_reraise_exception(): [ 565.598294] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.598294] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] self.force_reraise() [ 565.598294] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.598294] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] raise self.value [ 565.598294] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 565.598294] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] updated_port = self._update_port( [ 565.598294] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.598294] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] _ensure_no_port_binding_failure(port) [ 565.598294] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.598294] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] raise exception.PortBindingFailed(port_id=port['id']) [ 565.598647] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] nova.exception.PortBindingFailed: Binding failed for port 210424e1-1a59-4cf6-b39d-6ada30cc40e7, please check neutron logs for more information. [ 565.598647] env[61987]: ERROR nova.compute.manager [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] [ 565.598647] env[61987]: DEBUG nova.compute.utils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Binding failed for port 210424e1-1a59-4cf6-b39d-6ada30cc40e7, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 565.598647] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.247s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.603318] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Build of instance 1a9030ce-3a35-4760-96f4-4288c1d2f9d1 was re-scheduled: Binding failed for port 210424e1-1a59-4cf6-b39d-6ada30cc40e7, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 565.603318] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 565.603438] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquiring lock "refresh_cache-1a9030ce-3a35-4760-96f4-4288c1d2f9d1" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.603514] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquired lock "refresh_cache-1a9030ce-3a35-4760-96f4-4288c1d2f9d1" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.603676] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 566.142621] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 566.145033] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Acquiring lock "f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.145423] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Lock "f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.335032] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.503025] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eedcfaaa-d38c-4b4e-97c2-2a3078f27bed {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.512837] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1d5968b-7364-4fb5-a8ed-d2674b5440e9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.544598] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-177cf23b-91f9-49d1-9dc4-2d50afeea71e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.553274] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2519ac8-26a2-4fae-b524-64ffb76679e4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.568030] env[61987]: DEBUG nova.compute.provider_tree [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 566.709742] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Acquiring lock "6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.709970] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Lock "6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.839295] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Releasing lock "refresh_cache-1a9030ce-3a35-4760-96f4-4288c1d2f9d1" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.839295] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 566.839753] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 566.839753] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 566.867873] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 567.072567] env[61987]: DEBUG nova.scheduler.client.report [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 567.370635] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.501137] env[61987]: DEBUG oslo_concurrency.lockutils [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Acquiring lock "89a97f81-bcd9-49e5-bbd3-9140f6638a20" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.501137] env[61987]: DEBUG oslo_concurrency.lockutils [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Lock "89a97f81-bcd9-49e5-bbd3-9140f6638a20" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.578139] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.981s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.581675] env[61987]: ERROR nova.compute.manager [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 555557bc-6f57-400e-8dbb-216d588433f1, please check neutron logs for more information. [ 567.581675] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Traceback (most recent call last): [ 567.581675] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 567.581675] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] self.driver.spawn(context, instance, image_meta, [ 567.581675] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 567.581675] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 567.581675] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 567.581675] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] vm_ref = self.build_virtual_machine(instance, [ 567.581675] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 567.581675] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] vif_infos = vmwarevif.get_vif_info(self._session, [ 567.581675] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 567.582020] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] for vif in network_info: [ 567.582020] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 567.582020] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] return self._sync_wrapper(fn, *args, **kwargs) [ 567.582020] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 567.582020] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] self.wait() [ 567.582020] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 567.582020] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] self[:] = self._gt.wait() [ 567.582020] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 567.582020] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] return self._exit_event.wait() [ 567.582020] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 567.582020] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] result = hub.switch() [ 567.582020] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 567.582020] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] return self.greenlet.switch() [ 567.582390] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 567.582390] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] result = function(*args, **kwargs) [ 567.582390] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 567.582390] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] return func(*args, **kwargs) [ 567.582390] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 567.582390] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] raise e [ 567.582390] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 567.582390] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] nwinfo = self.network_api.allocate_for_instance( [ 567.582390] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 567.582390] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] created_port_ids = self._update_ports_for_instance( [ 567.582390] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 567.582390] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] with excutils.save_and_reraise_exception(): [ 567.582390] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 567.582881] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] self.force_reraise() [ 567.582881] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 567.582881] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] raise self.value [ 567.582881] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 567.582881] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] updated_port = self._update_port( [ 567.582881] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 567.582881] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] _ensure_no_port_binding_failure(port) [ 567.582881] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 567.582881] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] raise exception.PortBindingFailed(port_id=port['id']) [ 567.582881] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] nova.exception.PortBindingFailed: Binding failed for port 555557bc-6f57-400e-8dbb-216d588433f1, please check neutron logs for more information. [ 567.582881] env[61987]: ERROR nova.compute.manager [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] [ 567.583246] env[61987]: DEBUG nova.compute.utils [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Binding failed for port 555557bc-6f57-400e-8dbb-216d588433f1, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 567.584426] env[61987]: DEBUG nova.compute.manager [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Build of instance 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3 was re-scheduled: Binding failed for port 555557bc-6f57-400e-8dbb-216d588433f1, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 567.585561] env[61987]: DEBUG nova.compute.manager [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 567.587537] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "refresh_cache-9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.587821] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquired lock "refresh_cache-9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.588027] env[61987]: DEBUG nova.network.neutron [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 567.589545] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.676s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.874749] env[61987]: INFO nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 1a9030ce-3a35-4760-96f4-4288c1d2f9d1] Took 1.03 seconds to deallocate network for instance. [ 568.118933] env[61987]: DEBUG nova.network.neutron [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 568.308040] env[61987]: DEBUG nova.network.neutron [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.512016] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bbc7662-2efd-4e65-841b-121914e9aec3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.521161] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2467f46a-6435-4851-8283-6fd7b8023301 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.555695] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fc6e0ab-6add-455c-ad3c-5586406be6d0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.564696] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa46a5c9-6b76-4d3e-ad6f-ea0e938eac92 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.579233] env[61987]: DEBUG nova.compute.provider_tree [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 568.814484] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Releasing lock "refresh_cache-9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.814686] env[61987]: DEBUG nova.compute.manager [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 568.814905] env[61987]: DEBUG nova.compute.manager [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 568.815108] env[61987]: DEBUG nova.network.neutron [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 568.839200] env[61987]: DEBUG nova.network.neutron [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 568.915843] env[61987]: INFO nova.scheduler.client.report [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Deleted allocations for instance 1a9030ce-3a35-4760-96f4-4288c1d2f9d1 [ 569.085807] env[61987]: DEBUG nova.scheduler.client.report [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 569.342053] env[61987]: DEBUG nova.network.neutron [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.427545] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "1a9030ce-3a35-4760-96f4-4288c1d2f9d1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.035s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.592273] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.003s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.593220] env[61987]: ERROR nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5de8b27a-666f-4c53-9ef8-4b92e5c5a592, please check neutron logs for more information. [ 569.593220] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Traceback (most recent call last): [ 569.593220] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 569.593220] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] self.driver.spawn(context, instance, image_meta, [ 569.593220] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 569.593220] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] self._vmops.spawn(context, instance, image_meta, injected_files, [ 569.593220] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 569.593220] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] vm_ref = self.build_virtual_machine(instance, [ 569.593220] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 569.593220] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] vif_infos = vmwarevif.get_vif_info(self._session, [ 569.593220] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 569.593746] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] for vif in network_info: [ 569.593746] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 569.593746] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] return self._sync_wrapper(fn, *args, **kwargs) [ 569.593746] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 569.593746] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] self.wait() [ 569.593746] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 569.593746] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] self[:] = self._gt.wait() [ 569.593746] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 569.593746] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] return self._exit_event.wait() [ 569.593746] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 569.593746] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] result = hub.switch() [ 569.593746] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 569.593746] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] return self.greenlet.switch() [ 569.594427] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.594427] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] result = function(*args, **kwargs) [ 569.594427] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 569.594427] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] return func(*args, **kwargs) [ 569.594427] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 569.594427] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] raise e [ 569.594427] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 569.594427] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] nwinfo = self.network_api.allocate_for_instance( [ 569.594427] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 569.594427] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] created_port_ids = self._update_ports_for_instance( [ 569.594427] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 569.594427] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] with excutils.save_and_reraise_exception(): [ 569.594427] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.594772] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] self.force_reraise() [ 569.594772] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.594772] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] raise self.value [ 569.594772] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 569.594772] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] updated_port = self._update_port( [ 569.594772] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.594772] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] _ensure_no_port_binding_failure(port) [ 569.594772] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.594772] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] raise exception.PortBindingFailed(port_id=port['id']) [ 569.594772] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] nova.exception.PortBindingFailed: Binding failed for port 5de8b27a-666f-4c53-9ef8-4b92e5c5a592, please check neutron logs for more information. [ 569.594772] env[61987]: ERROR nova.compute.manager [instance: 510803e9-89ce-4452-b245-ee1f124ef081] [ 569.596193] env[61987]: DEBUG nova.compute.utils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Binding failed for port 5de8b27a-666f-4c53-9ef8-4b92e5c5a592, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 569.597480] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.451s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.598903] env[61987]: INFO nova.compute.claims [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 569.604242] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Build of instance 510803e9-89ce-4452-b245-ee1f124ef081 was re-scheduled: Binding failed for port 5de8b27a-666f-4c53-9ef8-4b92e5c5a592, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 569.606249] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 569.606249] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquiring lock "refresh_cache-510803e9-89ce-4452-b245-ee1f124ef081" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.606249] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquired lock "refresh_cache-510803e9-89ce-4452-b245-ee1f124ef081" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.606249] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 569.847382] env[61987]: INFO nova.compute.manager [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3] Took 1.03 seconds to deallocate network for instance. [ 569.930994] env[61987]: DEBUG nova.compute.manager [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 570.139083] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 570.324522] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.460703] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.829238] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Releasing lock "refresh_cache-510803e9-89ce-4452-b245-ee1f124ef081" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.829238] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 570.829238] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 570.829238] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 570.846034] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 570.879356] env[61987]: INFO nova.scheduler.client.report [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Deleted allocations for instance 9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3 [ 571.039681] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb99ef99-6263-472e-9102-2f115387a14b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.050450] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f223614-370d-4552-949f-4dc4da7de63f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.083445] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69cee3ae-016c-4824-a863-aa7c2e87231e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.091350] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-610adbe4-3534-447f-b764-563917e2c59b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.109475] env[61987]: DEBUG nova.compute.provider_tree [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.352780] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 571.393587] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d4c9dfb0-ab34-4657-bb91-9227fb521410 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "9eed7cf7-5911-4e21-b8c8-f5d90b0c5bc3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.889s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.612296] env[61987]: DEBUG oslo_concurrency.lockutils [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Acquiring lock "630e0959-6d7f-401d-b84c-1c098f54bf60" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.612696] env[61987]: DEBUG oslo_concurrency.lockutils [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Lock "630e0959-6d7f-401d-b84c-1c098f54bf60" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.615101] env[61987]: DEBUG nova.scheduler.client.report [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 571.856367] env[61987]: INFO nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: 510803e9-89ce-4452-b245-ee1f124ef081] Took 1.03 seconds to deallocate network for instance. [ 571.896675] env[61987]: DEBUG nova.compute.manager [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 572.124499] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.526s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.124966] env[61987]: DEBUG nova.compute.manager [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 572.130308] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.206s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.136347] env[61987]: INFO nova.compute.claims [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 572.426220] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.646637] env[61987]: DEBUG nova.compute.utils [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 572.649107] env[61987]: DEBUG nova.compute.manager [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 572.652454] env[61987]: DEBUG nova.network.neutron [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 572.695068] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Acquiring lock "466a5f5a-f48a-4921-8bcb-44a1840c8141" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.695068] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Lock "466a5f5a-f48a-4921-8bcb-44a1840c8141" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.742863] env[61987]: DEBUG nova.policy [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '15152ee0aa1f4d9db7e63ab08ad76a4b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2287d9f275945a5959acb3e6ef26074', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 572.900063] env[61987]: INFO nova.scheduler.client.report [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Deleted allocations for instance 510803e9-89ce-4452-b245-ee1f124ef081 [ 573.156459] env[61987]: DEBUG nova.compute.manager [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 573.372358] env[61987]: DEBUG nova.network.neutron [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Successfully created port: 36e6b6a6-fca0-41c9-8284-b06c55e23a01 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 573.411547] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "510803e9-89ce-4452-b245-ee1f124ef081" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.044s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.582409] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e30b961-948b-4fba-ae4f-4f9f8d3e5137 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.592901] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f267282-8e12-4055-a5d0-f1af32ef51bd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.628486] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-966d1fd5-2194-4b99-8af5-080bf1b83490 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.636260] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5298805e-3d06-4aee-b743-f656eff505c0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.650863] env[61987]: DEBUG nova.compute.provider_tree [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 573.914587] env[61987]: DEBUG nova.compute.manager [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 573.969983] env[61987]: DEBUG nova.network.neutron [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Successfully created port: b22af483-f9a1-481a-87e2-a9154ba4a4e2 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 574.157415] env[61987]: DEBUG nova.scheduler.client.report [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 574.169647] env[61987]: DEBUG nova.compute.manager [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 574.202785] env[61987]: DEBUG nova.virt.hardware [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 574.202785] env[61987]: DEBUG nova.virt.hardware [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 574.202785] env[61987]: DEBUG nova.virt.hardware [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 574.202785] env[61987]: DEBUG nova.virt.hardware [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 574.203050] env[61987]: DEBUG nova.virt.hardware [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 574.203050] env[61987]: DEBUG nova.virt.hardware [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 574.203954] env[61987]: DEBUG nova.virt.hardware [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 574.204273] env[61987]: DEBUG nova.virt.hardware [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 574.204556] env[61987]: DEBUG nova.virt.hardware [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 574.204829] env[61987]: DEBUG nova.virt.hardware [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 574.205144] env[61987]: DEBUG nova.virt.hardware [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 574.209157] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67339e85-c23c-4a0a-943b-16dc75fe3554 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.216964] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e9194b-e453-4e12-9d9f-dff9755e45bc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.262637] env[61987]: DEBUG oslo_concurrency.lockutils [None req-efff1759-dc77-4623-a6b0-5b38908831bd tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquiring lock "c9ce2911-8fc4-4aa8-883e-487ee06670ba" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.387523] env[61987]: DEBUG nova.network.neutron [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Successfully created port: 19feb53f-53ae-47ce-a5a3-8ec9f6c3b40f {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 574.446802] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.504636] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "ea6c2120-a69a-42c7-bc92-3e74539c008d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.504866] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "ea6c2120-a69a-42c7-bc92-3e74539c008d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.667038] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.538s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.667640] env[61987]: DEBUG nova.compute.manager [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 574.670349] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.475s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.008302] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "02fe490b-75c8-406b-98e0-aa200f249673" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.010940] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "02fe490b-75c8-406b-98e0-aa200f249673" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.176257] env[61987]: DEBUG nova.compute.utils [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 575.182373] env[61987]: DEBUG nova.compute.manager [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 575.182550] env[61987]: DEBUG nova.network.neutron [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 575.526200] env[61987]: DEBUG nova.policy [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6b9b9da9b9248d3bffdb97582f0b1fe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd4e086581d9e41b197f84dac49aab72e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 575.625181] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9617462d-7d91-40df-bd5e-eed2c04c96c0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.633453] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc9b83ad-5d97-40a9-9086-e17a90444ce9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.667149] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-257b00ce-a1d0-498d-ba43-3c3d5f5801f5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.675508] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e0a2044-37db-486d-9996-129e61561a41 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.697624] env[61987]: DEBUG nova.compute.manager [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 575.702116] env[61987]: DEBUG nova.compute.provider_tree [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 576.210525] env[61987]: DEBUG nova.scheduler.client.report [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 576.666111] env[61987]: DEBUG nova.network.neutron [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Successfully created port: 94295cbb-b5de-4392-853a-6e75332a2ac4 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 576.718313] env[61987]: DEBUG nova.compute.manager [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 576.722030] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.051s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.722261] env[61987]: ERROR nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 30ce6103-be54-438d-930c-51e75dbcf857, please check neutron logs for more information. [ 576.722261] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Traceback (most recent call last): [ 576.722261] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 576.722261] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] self.driver.spawn(context, instance, image_meta, [ 576.722261] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 576.722261] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] self._vmops.spawn(context, instance, image_meta, injected_files, [ 576.722261] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 576.722261] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] vm_ref = self.build_virtual_machine(instance, [ 576.722261] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 576.722261] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] vif_infos = vmwarevif.get_vif_info(self._session, [ 576.722261] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 576.722687] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] for vif in network_info: [ 576.722687] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 576.722687] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] return self._sync_wrapper(fn, *args, **kwargs) [ 576.722687] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 576.722687] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] self.wait() [ 576.722687] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 576.722687] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] self[:] = self._gt.wait() [ 576.722687] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 576.722687] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] return self._exit_event.wait() [ 576.722687] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 576.722687] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] result = hub.switch() [ 576.722687] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 576.722687] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] return self.greenlet.switch() [ 576.723156] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 576.723156] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] result = function(*args, **kwargs) [ 576.723156] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 576.723156] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] return func(*args, **kwargs) [ 576.723156] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 576.723156] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] raise e [ 576.723156] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 576.723156] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] nwinfo = self.network_api.allocate_for_instance( [ 576.723156] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 576.723156] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] created_port_ids = self._update_ports_for_instance( [ 576.723156] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 576.723156] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] with excutils.save_and_reraise_exception(): [ 576.723156] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.723974] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] self.force_reraise() [ 576.723974] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.723974] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] raise self.value [ 576.723974] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 576.723974] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] updated_port = self._update_port( [ 576.723974] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.723974] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] _ensure_no_port_binding_failure(port) [ 576.723974] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.723974] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] raise exception.PortBindingFailed(port_id=port['id']) [ 576.723974] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] nova.exception.PortBindingFailed: Binding failed for port 30ce6103-be54-438d-930c-51e75dbcf857, please check neutron logs for more information. [ 576.723974] env[61987]: ERROR nova.compute.manager [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] [ 576.724420] env[61987]: DEBUG nova.compute.utils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Binding failed for port 30ce6103-be54-438d-930c-51e75dbcf857, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 576.726461] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.034s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.729398] env[61987]: INFO nova.compute.claims [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 576.732838] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Build of instance c9ce2911-8fc4-4aa8-883e-487ee06670ba was re-scheduled: Binding failed for port 30ce6103-be54-438d-930c-51e75dbcf857, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 576.733706] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 576.733950] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquiring lock "refresh_cache-c9ce2911-8fc4-4aa8-883e-487ee06670ba" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.734207] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquired lock "refresh_cache-c9ce2911-8fc4-4aa8-883e-487ee06670ba" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.734415] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 576.757561] env[61987]: DEBUG nova.virt.hardware [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 576.757561] env[61987]: DEBUG nova.virt.hardware [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 576.757561] env[61987]: DEBUG nova.virt.hardware [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 576.757561] env[61987]: DEBUG nova.virt.hardware [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 576.757769] env[61987]: DEBUG nova.virt.hardware [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 576.757848] env[61987]: DEBUG nova.virt.hardware [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 576.760513] env[61987]: DEBUG nova.virt.hardware [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 576.760656] env[61987]: DEBUG nova.virt.hardware [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 576.760838] env[61987]: DEBUG nova.virt.hardware [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 576.761764] env[61987]: DEBUG nova.virt.hardware [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 576.762041] env[61987]: DEBUG nova.virt.hardware [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 576.763333] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e67ad642-39ed-451c-8930-f7c667bf8d64 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.773540] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77cc9a94-7acd-4a3d-a9ac-02fc1c40a7d7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.201995] env[61987]: ERROR nova.compute.manager [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 36e6b6a6-fca0-41c9-8284-b06c55e23a01, please check neutron logs for more information. [ 577.201995] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 577.201995] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 577.201995] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 577.201995] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.201995] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 577.201995] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.201995] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 577.201995] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.201995] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 577.201995] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.201995] env[61987]: ERROR nova.compute.manager raise self.value [ 577.201995] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.201995] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 577.201995] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.201995] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 577.202590] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.202590] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 577.202590] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 36e6b6a6-fca0-41c9-8284-b06c55e23a01, please check neutron logs for more information. [ 577.202590] env[61987]: ERROR nova.compute.manager [ 577.202590] env[61987]: Traceback (most recent call last): [ 577.202590] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 577.202590] env[61987]: listener.cb(fileno) [ 577.202590] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.202590] env[61987]: result = function(*args, **kwargs) [ 577.202590] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 577.202590] env[61987]: return func(*args, **kwargs) [ 577.202590] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 577.202590] env[61987]: raise e [ 577.202590] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 577.202590] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 577.202590] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.202590] env[61987]: created_port_ids = self._update_ports_for_instance( [ 577.202590] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.202590] env[61987]: with excutils.save_and_reraise_exception(): [ 577.202590] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.202590] env[61987]: self.force_reraise() [ 577.202590] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.202590] env[61987]: raise self.value [ 577.202590] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.202590] env[61987]: updated_port = self._update_port( [ 577.202590] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.202590] env[61987]: _ensure_no_port_binding_failure(port) [ 577.202590] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.202590] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 577.203484] env[61987]: nova.exception.PortBindingFailed: Binding failed for port 36e6b6a6-fca0-41c9-8284-b06c55e23a01, please check neutron logs for more information. [ 577.203484] env[61987]: Removing descriptor: 19 [ 577.203484] env[61987]: ERROR nova.compute.manager [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 36e6b6a6-fca0-41c9-8284-b06c55e23a01, please check neutron logs for more information. [ 577.203484] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Traceback (most recent call last): [ 577.203484] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 577.203484] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] yield resources [ 577.203484] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 577.203484] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] self.driver.spawn(context, instance, image_meta, [ 577.203484] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 577.203484] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 577.203484] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 577.203484] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] vm_ref = self.build_virtual_machine(instance, [ 577.203839] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 577.203839] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] vif_infos = vmwarevif.get_vif_info(self._session, [ 577.203839] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 577.203839] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] for vif in network_info: [ 577.203839] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 577.203839] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] return self._sync_wrapper(fn, *args, **kwargs) [ 577.203839] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 577.203839] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] self.wait() [ 577.203839] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 577.203839] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] self[:] = self._gt.wait() [ 577.203839] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 577.203839] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] return self._exit_event.wait() [ 577.203839] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 577.204254] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] result = hub.switch() [ 577.204254] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 577.204254] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] return self.greenlet.switch() [ 577.204254] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.204254] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] result = function(*args, **kwargs) [ 577.204254] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 577.204254] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] return func(*args, **kwargs) [ 577.204254] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 577.204254] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] raise e [ 577.204254] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 577.204254] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] nwinfo = self.network_api.allocate_for_instance( [ 577.204254] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 577.204254] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] created_port_ids = self._update_ports_for_instance( [ 577.204679] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 577.204679] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] with excutils.save_and_reraise_exception(): [ 577.204679] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.204679] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] self.force_reraise() [ 577.204679] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.204679] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] raise self.value [ 577.204679] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 577.204679] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] updated_port = self._update_port( [ 577.204679] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.204679] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] _ensure_no_port_binding_failure(port) [ 577.204679] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.204679] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] raise exception.PortBindingFailed(port_id=port['id']) [ 577.205211] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] nova.exception.PortBindingFailed: Binding failed for port 36e6b6a6-fca0-41c9-8284-b06c55e23a01, please check neutron logs for more information. [ 577.205211] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] [ 577.205211] env[61987]: INFO nova.compute.manager [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Terminating instance [ 577.290668] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 577.353107] env[61987]: DEBUG nova.compute.manager [req-a62a3923-49bc-441e-b984-a1dae51943eb req-5d368f85-b1ac-4fe1-9624-0d896fe3a0ee service nova] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Received event network-changed-36e6b6a6-fca0-41c9-8284-b06c55e23a01 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 577.353329] env[61987]: DEBUG nova.compute.manager [req-a62a3923-49bc-441e-b984-a1dae51943eb req-5d368f85-b1ac-4fe1-9624-0d896fe3a0ee service nova] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Refreshing instance network info cache due to event network-changed-36e6b6a6-fca0-41c9-8284-b06c55e23a01. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 577.353531] env[61987]: DEBUG oslo_concurrency.lockutils [req-a62a3923-49bc-441e-b984-a1dae51943eb req-5d368f85-b1ac-4fe1-9624-0d896fe3a0ee service nova] Acquiring lock "refresh_cache-7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.353667] env[61987]: DEBUG oslo_concurrency.lockutils [req-a62a3923-49bc-441e-b984-a1dae51943eb req-5d368f85-b1ac-4fe1-9624-0d896fe3a0ee service nova] Acquired lock "refresh_cache-7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.353875] env[61987]: DEBUG nova.network.neutron [req-a62a3923-49bc-441e-b984-a1dae51943eb req-5d368f85-b1ac-4fe1-9624-0d896fe3a0ee service nova] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Refreshing network info cache for port 36e6b6a6-fca0-41c9-8284-b06c55e23a01 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 577.482940] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.721872] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Acquiring lock "refresh_cache-7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.918335] env[61987]: DEBUG oslo_concurrency.lockutils [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Acquiring lock "0b4b3bdb-7393-47ab-abc1-44c47496cb3a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.918335] env[61987]: DEBUG oslo_concurrency.lockutils [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Lock "0b4b3bdb-7393-47ab-abc1-44c47496cb3a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.933860] env[61987]: DEBUG nova.network.neutron [req-a62a3923-49bc-441e-b984-a1dae51943eb req-5d368f85-b1ac-4fe1-9624-0d896fe3a0ee service nova] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 577.986913] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Releasing lock "refresh_cache-c9ce2911-8fc4-4aa8-883e-487ee06670ba" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.987181] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 577.987370] env[61987]: DEBUG nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 577.987551] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 578.013384] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 578.279375] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5a57454-df8c-4817-881e-f5dd17664825 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.295346] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5414c642-ae08-4291-90cf-4904b2ad41e0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.336436] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a98d3600-745c-4581-8663-402a45c633d8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.347561] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-633ad627-018f-49eb-b7b0-c700c7d6bfa4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.365257] env[61987]: DEBUG nova.compute.provider_tree [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 578.433317] env[61987]: DEBUG nova.network.neutron [req-a62a3923-49bc-441e-b984-a1dae51943eb req-5d368f85-b1ac-4fe1-9624-0d896fe3a0ee service nova] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.520396] env[61987]: DEBUG nova.network.neutron [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.868646] env[61987]: DEBUG nova.scheduler.client.report [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 578.940633] env[61987]: DEBUG oslo_concurrency.lockutils [req-a62a3923-49bc-441e-b984-a1dae51943eb req-5d368f85-b1ac-4fe1-9624-0d896fe3a0ee service nova] Releasing lock "refresh_cache-7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.940633] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Acquired lock "refresh_cache-7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.940633] env[61987]: DEBUG nova.network.neutron [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 579.006383] env[61987]: ERROR nova.compute.manager [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 94295cbb-b5de-4392-853a-6e75332a2ac4, please check neutron logs for more information. [ 579.006383] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 579.006383] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 579.006383] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 579.006383] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 579.006383] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 579.006383] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 579.006383] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 579.006383] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.006383] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 579.006383] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.006383] env[61987]: ERROR nova.compute.manager raise self.value [ 579.006383] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 579.006383] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 579.006383] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.006383] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 579.006902] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.006902] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 579.006902] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 94295cbb-b5de-4392-853a-6e75332a2ac4, please check neutron logs for more information. [ 579.006902] env[61987]: ERROR nova.compute.manager [ 579.006902] env[61987]: Traceback (most recent call last): [ 579.006902] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 579.006902] env[61987]: listener.cb(fileno) [ 579.006902] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.006902] env[61987]: result = function(*args, **kwargs) [ 579.006902] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 579.006902] env[61987]: return func(*args, **kwargs) [ 579.006902] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 579.006902] env[61987]: raise e [ 579.006902] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 579.006902] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 579.006902] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 579.006902] env[61987]: created_port_ids = self._update_ports_for_instance( [ 579.006902] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 579.006902] env[61987]: with excutils.save_and_reraise_exception(): [ 579.006902] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.006902] env[61987]: self.force_reraise() [ 579.006902] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.006902] env[61987]: raise self.value [ 579.006902] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 579.006902] env[61987]: updated_port = self._update_port( [ 579.006902] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.006902] env[61987]: _ensure_no_port_binding_failure(port) [ 579.006902] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.006902] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 579.007704] env[61987]: nova.exception.PortBindingFailed: Binding failed for port 94295cbb-b5de-4392-853a-6e75332a2ac4, please check neutron logs for more information. [ 579.007704] env[61987]: Removing descriptor: 18 [ 579.007704] env[61987]: ERROR nova.compute.manager [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 94295cbb-b5de-4392-853a-6e75332a2ac4, please check neutron logs for more information. [ 579.007704] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Traceback (most recent call last): [ 579.007704] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 579.007704] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] yield resources [ 579.007704] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 579.007704] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] self.driver.spawn(context, instance, image_meta, [ 579.007704] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 579.007704] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] self._vmops.spawn(context, instance, image_meta, injected_files, [ 579.007704] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 579.007704] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] vm_ref = self.build_virtual_machine(instance, [ 579.008108] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 579.008108] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] vif_infos = vmwarevif.get_vif_info(self._session, [ 579.008108] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 579.008108] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] for vif in network_info: [ 579.008108] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 579.008108] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] return self._sync_wrapper(fn, *args, **kwargs) [ 579.008108] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 579.008108] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] self.wait() [ 579.008108] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 579.008108] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] self[:] = self._gt.wait() [ 579.008108] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 579.008108] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] return self._exit_event.wait() [ 579.008108] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 579.008472] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] result = hub.switch() [ 579.008472] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 579.008472] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] return self.greenlet.switch() [ 579.008472] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.008472] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] result = function(*args, **kwargs) [ 579.008472] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 579.008472] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] return func(*args, **kwargs) [ 579.008472] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 579.008472] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] raise e [ 579.008472] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 579.008472] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] nwinfo = self.network_api.allocate_for_instance( [ 579.008472] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 579.008472] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] created_port_ids = self._update_ports_for_instance( [ 579.008834] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 579.008834] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] with excutils.save_and_reraise_exception(): [ 579.008834] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.008834] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] self.force_reraise() [ 579.008834] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.008834] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] raise self.value [ 579.008834] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 579.008834] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] updated_port = self._update_port( [ 579.008834] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.008834] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] _ensure_no_port_binding_failure(port) [ 579.008834] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.008834] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] raise exception.PortBindingFailed(port_id=port['id']) [ 579.009180] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] nova.exception.PortBindingFailed: Binding failed for port 94295cbb-b5de-4392-853a-6e75332a2ac4, please check neutron logs for more information. [ 579.009180] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] [ 579.009180] env[61987]: INFO nova.compute.manager [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Terminating instance [ 579.027160] env[61987]: INFO nova.compute.manager [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Took 1.04 seconds to deallocate network for instance. [ 579.373307] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.647s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.373965] env[61987]: DEBUG nova.compute.manager [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 579.380493] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.635s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.381722] env[61987]: INFO nova.compute.claims [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 579.507939] env[61987]: DEBUG nova.network.neutron [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 579.513869] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquiring lock "refresh_cache-3182ca24-c989-4c29-8e47-d913d0607240" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.513869] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquired lock "refresh_cache-3182ca24-c989-4c29-8e47-d913d0607240" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.513869] env[61987]: DEBUG nova.network.neutron [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 579.637658] env[61987]: DEBUG nova.network.neutron [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.809795] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "51a36102-795e-47b7-b96a-857e54dc703e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.810035] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "51a36102-795e-47b7-b96a-857e54dc703e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.827899] env[61987]: DEBUG nova.compute.manager [req-2e835080-1822-4f57-8573-8416ecec98dc req-a8e0f5af-82d2-4b99-96f1-98c478cb06d2 service nova] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Received event network-vif-deleted-36e6b6a6-fca0-41c9-8284-b06c55e23a01 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 579.828184] env[61987]: DEBUG nova.compute.manager [req-2e835080-1822-4f57-8573-8416ecec98dc req-a8e0f5af-82d2-4b99-96f1-98c478cb06d2 service nova] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Received event network-changed-94295cbb-b5de-4392-853a-6e75332a2ac4 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 579.828265] env[61987]: DEBUG nova.compute.manager [req-2e835080-1822-4f57-8573-8416ecec98dc req-a8e0f5af-82d2-4b99-96f1-98c478cb06d2 service nova] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Refreshing instance network info cache due to event network-changed-94295cbb-b5de-4392-853a-6e75332a2ac4. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 579.828447] env[61987]: DEBUG oslo_concurrency.lockutils [req-2e835080-1822-4f57-8573-8416ecec98dc req-a8e0f5af-82d2-4b99-96f1-98c478cb06d2 service nova] Acquiring lock "refresh_cache-3182ca24-c989-4c29-8e47-d913d0607240" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.883016] env[61987]: DEBUG nova.compute.utils [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 579.884525] env[61987]: DEBUG nova.compute.manager [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 579.885185] env[61987]: DEBUG nova.network.neutron [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 579.948019] env[61987]: DEBUG nova.policy [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '92795dce5ff54f00952ee3181ca83b2b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '676b4523ecc44a47b80109d50b018b34', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 580.038249] env[61987]: DEBUG nova.network.neutron [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.067337] env[61987]: INFO nova.scheduler.client.report [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Deleted allocations for instance c9ce2911-8fc4-4aa8-883e-487ee06670ba [ 580.143023] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Releasing lock "refresh_cache-7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.143023] env[61987]: DEBUG nova.compute.manager [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 580.143023] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 580.143023] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ed23388c-590c-4f47-a98c-ab4c2510c7ce {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.154492] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4cddba1-929f-44b1-af94-3a8c8b8a8d7f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.166196] env[61987]: DEBUG nova.network.neutron [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.179582] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7 could not be found. [ 580.179745] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 580.179891] env[61987]: INFO nova.compute.manager [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 580.180153] env[61987]: DEBUG oslo.service.loopingcall [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 580.180610] env[61987]: DEBUG nova.compute.manager [-] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 580.180610] env[61987]: DEBUG nova.network.neutron [-] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 580.231381] env[61987]: DEBUG nova.network.neutron [-] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.298722] env[61987]: DEBUG oslo_concurrency.lockutils [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Acquiring lock "fa943137-caf1-4b41-b644-c9146f72f956" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.298970] env[61987]: DEBUG oslo_concurrency.lockutils [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Lock "fa943137-caf1-4b41-b644-c9146f72f956" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.392574] env[61987]: DEBUG nova.compute.manager [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 580.447895] env[61987]: DEBUG nova.network.neutron [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Successfully created port: 53be8281-4dc3-4ab6-abb1-a65876f596ef {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 580.582027] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b71be046-372b-486f-98b5-9c345016d0b4 tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "c9ce2911-8fc4-4aa8-883e-487ee06670ba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.139s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.589478] env[61987]: DEBUG oslo_concurrency.lockutils [None req-efff1759-dc77-4623-a6b0-5b38908831bd tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "c9ce2911-8fc4-4aa8-883e-487ee06670ba" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 6.322s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.589478] env[61987]: DEBUG oslo_concurrency.lockutils [None req-efff1759-dc77-4623-a6b0-5b38908831bd tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquiring lock "c9ce2911-8fc4-4aa8-883e-487ee06670ba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.589478] env[61987]: DEBUG oslo_concurrency.lockutils [None req-efff1759-dc77-4623-a6b0-5b38908831bd tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "c9ce2911-8fc4-4aa8-883e-487ee06670ba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.589478] env[61987]: DEBUG oslo_concurrency.lockutils [None req-efff1759-dc77-4623-a6b0-5b38908831bd tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "c9ce2911-8fc4-4aa8-883e-487ee06670ba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.595064] env[61987]: INFO nova.compute.manager [None req-efff1759-dc77-4623-a6b0-5b38908831bd tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Terminating instance [ 580.672210] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Releasing lock "refresh_cache-3182ca24-c989-4c29-8e47-d913d0607240" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.672210] env[61987]: DEBUG nova.compute.manager [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 580.672210] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 580.672839] env[61987]: DEBUG oslo_concurrency.lockutils [req-2e835080-1822-4f57-8573-8416ecec98dc req-a8e0f5af-82d2-4b99-96f1-98c478cb06d2 service nova] Acquired lock "refresh_cache-3182ca24-c989-4c29-8e47-d913d0607240" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.674024] env[61987]: DEBUG nova.network.neutron [req-2e835080-1822-4f57-8573-8416ecec98dc req-a8e0f5af-82d2-4b99-96f1-98c478cb06d2 service nova] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Refreshing network info cache for port 94295cbb-b5de-4392-853a-6e75332a2ac4 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 580.674794] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fac9ebaf-0bf3-4505-b50c-7df8a02bf070 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.690862] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c296738f-5345-4a61-bc0d-4515381a9cef {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.715604] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3182ca24-c989-4c29-8e47-d913d0607240 could not be found. [ 580.715843] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 580.716362] env[61987]: INFO nova.compute.manager [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Took 0.04 seconds to destroy the instance on the hypervisor. [ 580.716687] env[61987]: DEBUG oslo.service.loopingcall [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 580.719250] env[61987]: DEBUG nova.compute.manager [-] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 580.719359] env[61987]: DEBUG nova.network.neutron [-] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 580.750078] env[61987]: DEBUG nova.network.neutron [-] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.908154] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a81b2cb-7bf9-452b-933f-c61ffde5072c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.918307] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3934419-bace-4407-bf65-5527255ba70a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.956158] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f76139d-7286-4320-b04f-0e48565b58c7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.967796] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-971a7728-435e-43e2-b9e9-73982452e86d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.985925] env[61987]: DEBUG nova.compute.provider_tree [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.092438] env[61987]: DEBUG nova.compute.manager [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 581.103583] env[61987]: DEBUG oslo_concurrency.lockutils [None req-efff1759-dc77-4623-a6b0-5b38908831bd tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquiring lock "refresh_cache-c9ce2911-8fc4-4aa8-883e-487ee06670ba" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.103767] env[61987]: DEBUG oslo_concurrency.lockutils [None req-efff1759-dc77-4623-a6b0-5b38908831bd tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Acquired lock "refresh_cache-c9ce2911-8fc4-4aa8-883e-487ee06670ba" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.103980] env[61987]: DEBUG nova.network.neutron [None req-efff1759-dc77-4623-a6b0-5b38908831bd tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 581.212869] env[61987]: DEBUG nova.network.neutron [req-2e835080-1822-4f57-8573-8416ecec98dc req-a8e0f5af-82d2-4b99-96f1-98c478cb06d2 service nova] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.258338] env[61987]: DEBUG nova.network.neutron [-] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.297572] env[61987]: DEBUG nova.network.neutron [req-2e835080-1822-4f57-8573-8416ecec98dc req-a8e0f5af-82d2-4b99-96f1-98c478cb06d2 service nova] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.407826] env[61987]: DEBUG nova.compute.manager [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 581.445957] env[61987]: DEBUG nova.virt.hardware [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 581.446223] env[61987]: DEBUG nova.virt.hardware [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 581.446379] env[61987]: DEBUG nova.virt.hardware [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 581.446560] env[61987]: DEBUG nova.virt.hardware [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 581.446705] env[61987]: DEBUG nova.virt.hardware [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 581.446859] env[61987]: DEBUG nova.virt.hardware [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 581.447790] env[61987]: DEBUG nova.virt.hardware [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 581.447790] env[61987]: DEBUG nova.virt.hardware [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 581.447790] env[61987]: DEBUG nova.virt.hardware [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 581.447790] env[61987]: DEBUG nova.virt.hardware [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 581.448139] env[61987]: DEBUG nova.virt.hardware [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 581.448899] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83885941-6b92-40c6-8f7d-32a1eac1f0ab {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.457656] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e315d2b4-f2af-43d8-9435-6ee74e5429e2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.488758] env[61987]: DEBUG nova.scheduler.client.report [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 581.512104] env[61987]: DEBUG nova.network.neutron [-] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.622392] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.642167] env[61987]: DEBUG nova.network.neutron [None req-efff1759-dc77-4623-a6b0-5b38908831bd tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.761457] env[61987]: INFO nova.compute.manager [-] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Took 1.04 seconds to deallocate network for instance. [ 581.765016] env[61987]: DEBUG nova.compute.claims [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 581.765016] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.801206] env[61987]: DEBUG oslo_concurrency.lockutils [req-2e835080-1822-4f57-8573-8416ecec98dc req-a8e0f5af-82d2-4b99-96f1-98c478cb06d2 service nova] Releasing lock "refresh_cache-3182ca24-c989-4c29-8e47-d913d0607240" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.801655] env[61987]: DEBUG nova.compute.manager [req-2e835080-1822-4f57-8573-8416ecec98dc req-a8e0f5af-82d2-4b99-96f1-98c478cb06d2 service nova] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Received event network-vif-deleted-94295cbb-b5de-4392-853a-6e75332a2ac4 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 581.821371] env[61987]: DEBUG nova.network.neutron [None req-efff1759-dc77-4623-a6b0-5b38908831bd tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.996140] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.616s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.996592] env[61987]: DEBUG nova.compute.manager [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 582.001977] env[61987]: DEBUG oslo_concurrency.lockutils [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.485s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.004671] env[61987]: INFO nova.compute.claims [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 582.009691] env[61987]: DEBUG nova.compute.manager [req-8263620d-7a09-4ed2-98c1-adcc2338a4ef req-8bcc3b2f-043f-4a8a-8f39-196688402d0d service nova] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Received event network-changed-53be8281-4dc3-4ab6-abb1-a65876f596ef {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 582.009882] env[61987]: DEBUG nova.compute.manager [req-8263620d-7a09-4ed2-98c1-adcc2338a4ef req-8bcc3b2f-043f-4a8a-8f39-196688402d0d service nova] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Refreshing instance network info cache due to event network-changed-53be8281-4dc3-4ab6-abb1-a65876f596ef. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 582.010107] env[61987]: DEBUG oslo_concurrency.lockutils [req-8263620d-7a09-4ed2-98c1-adcc2338a4ef req-8bcc3b2f-043f-4a8a-8f39-196688402d0d service nova] Acquiring lock "refresh_cache-7a94f776-579d-41e8-94aa-6095b3fd69c6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.010244] env[61987]: DEBUG oslo_concurrency.lockutils [req-8263620d-7a09-4ed2-98c1-adcc2338a4ef req-8bcc3b2f-043f-4a8a-8f39-196688402d0d service nova] Acquired lock "refresh_cache-7a94f776-579d-41e8-94aa-6095b3fd69c6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.010397] env[61987]: DEBUG nova.network.neutron [req-8263620d-7a09-4ed2-98c1-adcc2338a4ef req-8bcc3b2f-043f-4a8a-8f39-196688402d0d service nova] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Refreshing network info cache for port 53be8281-4dc3-4ab6-abb1-a65876f596ef {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 582.015784] env[61987]: INFO nova.compute.manager [-] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Took 1.84 seconds to deallocate network for instance. [ 582.020100] env[61987]: DEBUG nova.compute.claims [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 582.020100] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.030706] env[61987]: ERROR nova.compute.manager [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 53be8281-4dc3-4ab6-abb1-a65876f596ef, please check neutron logs for more information. [ 582.030706] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 582.030706] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 582.030706] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 582.030706] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 582.030706] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 582.030706] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 582.030706] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 582.030706] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.030706] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 582.030706] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.030706] env[61987]: ERROR nova.compute.manager raise self.value [ 582.030706] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 582.030706] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 582.030706] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.030706] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 582.031267] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.031267] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 582.031267] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 53be8281-4dc3-4ab6-abb1-a65876f596ef, please check neutron logs for more information. [ 582.031267] env[61987]: ERROR nova.compute.manager [ 582.031267] env[61987]: Traceback (most recent call last): [ 582.031267] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 582.031267] env[61987]: listener.cb(fileno) [ 582.031267] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.031267] env[61987]: result = function(*args, **kwargs) [ 582.031267] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.031267] env[61987]: return func(*args, **kwargs) [ 582.031267] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 582.031267] env[61987]: raise e [ 582.031267] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 582.031267] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 582.031267] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 582.031267] env[61987]: created_port_ids = self._update_ports_for_instance( [ 582.031267] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 582.031267] env[61987]: with excutils.save_and_reraise_exception(): [ 582.031267] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.031267] env[61987]: self.force_reraise() [ 582.031267] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.031267] env[61987]: raise self.value [ 582.031267] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 582.031267] env[61987]: updated_port = self._update_port( [ 582.031267] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.031267] env[61987]: _ensure_no_port_binding_failure(port) [ 582.031267] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.031267] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 582.032313] env[61987]: nova.exception.PortBindingFailed: Binding failed for port 53be8281-4dc3-4ab6-abb1-a65876f596ef, please check neutron logs for more information. [ 582.032313] env[61987]: Removing descriptor: 18 [ 582.032313] env[61987]: ERROR nova.compute.manager [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 53be8281-4dc3-4ab6-abb1-a65876f596ef, please check neutron logs for more information. [ 582.032313] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Traceback (most recent call last): [ 582.032313] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 582.032313] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] yield resources [ 582.032313] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 582.032313] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] self.driver.spawn(context, instance, image_meta, [ 582.032313] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 582.032313] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 582.032313] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 582.032313] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] vm_ref = self.build_virtual_machine(instance, [ 582.032716] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 582.032716] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] vif_infos = vmwarevif.get_vif_info(self._session, [ 582.032716] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 582.032716] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] for vif in network_info: [ 582.032716] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 582.032716] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] return self._sync_wrapper(fn, *args, **kwargs) [ 582.032716] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 582.032716] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] self.wait() [ 582.032716] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 582.032716] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] self[:] = self._gt.wait() [ 582.032716] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 582.032716] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] return self._exit_event.wait() [ 582.032716] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 582.033248] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] result = hub.switch() [ 582.033248] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 582.033248] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] return self.greenlet.switch() [ 582.033248] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.033248] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] result = function(*args, **kwargs) [ 582.033248] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.033248] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] return func(*args, **kwargs) [ 582.033248] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 582.033248] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] raise e [ 582.033248] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 582.033248] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] nwinfo = self.network_api.allocate_for_instance( [ 582.033248] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 582.033248] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] created_port_ids = self._update_ports_for_instance( [ 582.035263] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 582.035263] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] with excutils.save_and_reraise_exception(): [ 582.035263] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.035263] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] self.force_reraise() [ 582.035263] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.035263] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] raise self.value [ 582.035263] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 582.035263] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] updated_port = self._update_port( [ 582.035263] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.035263] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] _ensure_no_port_binding_failure(port) [ 582.035263] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.035263] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] raise exception.PortBindingFailed(port_id=port['id']) [ 582.035637] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] nova.exception.PortBindingFailed: Binding failed for port 53be8281-4dc3-4ab6-abb1-a65876f596ef, please check neutron logs for more information. [ 582.035637] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] [ 582.035637] env[61987]: INFO nova.compute.manager [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Terminating instance [ 582.330585] env[61987]: DEBUG oslo_concurrency.lockutils [None req-efff1759-dc77-4623-a6b0-5b38908831bd tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Releasing lock "refresh_cache-c9ce2911-8fc4-4aa8-883e-487ee06670ba" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.331277] env[61987]: DEBUG nova.compute.manager [None req-efff1759-dc77-4623-a6b0-5b38908831bd tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 582.331659] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-efff1759-dc77-4623-a6b0-5b38908831bd tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 582.331978] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1c5017ff-216b-4a86-aea8-4327154416ee {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.345166] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-513a183b-0dfe-4f02-b3ff-4c613ac8fa92 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.377290] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-efff1759-dc77-4623-a6b0-5b38908831bd tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c9ce2911-8fc4-4aa8-883e-487ee06670ba could not be found. [ 582.377526] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-efff1759-dc77-4623-a6b0-5b38908831bd tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 582.377699] env[61987]: INFO nova.compute.manager [None req-efff1759-dc77-4623-a6b0-5b38908831bd tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Took 0.05 seconds to destroy the instance on the hypervisor. [ 582.377924] env[61987]: DEBUG oslo.service.loopingcall [None req-efff1759-dc77-4623-a6b0-5b38908831bd tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 582.378158] env[61987]: DEBUG nova.compute.manager [-] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 582.378252] env[61987]: DEBUG nova.network.neutron [-] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 582.397024] env[61987]: DEBUG nova.network.neutron [-] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 582.516152] env[61987]: DEBUG nova.compute.utils [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 582.521159] env[61987]: DEBUG nova.compute.manager [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 582.521159] env[61987]: DEBUG nova.network.neutron [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 582.542114] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Acquiring lock "refresh_cache-7a94f776-579d-41e8-94aa-6095b3fd69c6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.559869] env[61987]: DEBUG nova.network.neutron [req-8263620d-7a09-4ed2-98c1-adcc2338a4ef req-8bcc3b2f-043f-4a8a-8f39-196688402d0d service nova] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 582.656351] env[61987]: DEBUG nova.policy [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '42dbefbe684c45669b9ab89c7f636d5f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '32aa82b5b3f54c2280e062f4c8639274', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 582.810218] env[61987]: DEBUG nova.network.neutron [req-8263620d-7a09-4ed2-98c1-adcc2338a4ef req-8bcc3b2f-043f-4a8a-8f39-196688402d0d service nova] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.899587] env[61987]: DEBUG nova.network.neutron [-] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.912146] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Acquiring lock "49e0d752-6f88-45c4-bd6a-d9fe4f526a2f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.912146] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Lock "49e0d752-6f88-45c4-bd6a-d9fe4f526a2f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.024526] env[61987]: DEBUG nova.compute.manager [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 583.308476] env[61987]: DEBUG nova.network.neutron [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Successfully created port: 1aae5742-c738-447d-b60d-05f7c8f67871 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 583.313077] env[61987]: DEBUG oslo_concurrency.lockutils [req-8263620d-7a09-4ed2-98c1-adcc2338a4ef req-8bcc3b2f-043f-4a8a-8f39-196688402d0d service nova] Releasing lock "refresh_cache-7a94f776-579d-41e8-94aa-6095b3fd69c6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.314546] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Acquired lock "refresh_cache-7a94f776-579d-41e8-94aa-6095b3fd69c6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.314749] env[61987]: DEBUG nova.network.neutron [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 583.409520] env[61987]: INFO nova.compute.manager [-] [instance: c9ce2911-8fc4-4aa8-883e-487ee06670ba] Took 1.03 seconds to deallocate network for instance. [ 583.501359] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Acquiring lock "8ff4ebb9-57ea-49d2-8d3c-5876da61e77b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.501815] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Lock "8ff4ebb9-57ea-49d2-8d3c-5876da61e77b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.564329] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ce32def-9923-4706-937a-08742a9a1141 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.574886] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f0bff45-0255-47e8-aee0-750b76bca4fe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.606754] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81be3a28-bd79-4cb2-b3de-c24401a1a868 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.614459] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e0a57ff-399e-45da-912e-10efaab29932 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.629199] env[61987]: DEBUG nova.compute.provider_tree [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 583.839664] env[61987]: DEBUG nova.network.neutron [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 583.983659] env[61987]: DEBUG nova.network.neutron [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.042553] env[61987]: DEBUG nova.compute.manager [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 584.065712] env[61987]: DEBUG nova.compute.manager [req-5731c0ca-8ca3-4b49-bac4-cbaa3664b29c req-22a9ef46-4aad-48fb-be49-459a08a8a1eb service nova] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Received event network-vif-deleted-53be8281-4dc3-4ab6-abb1-a65876f596ef {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 584.082311] env[61987]: DEBUG nova.virt.hardware [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 584.082311] env[61987]: DEBUG nova.virt.hardware [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 584.082516] env[61987]: DEBUG nova.virt.hardware [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 584.083883] env[61987]: DEBUG nova.virt.hardware [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 584.083883] env[61987]: DEBUG nova.virt.hardware [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 584.083883] env[61987]: DEBUG nova.virt.hardware [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 584.088105] env[61987]: DEBUG nova.virt.hardware [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 584.088105] env[61987]: DEBUG nova.virt.hardware [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 584.088105] env[61987]: DEBUG nova.virt.hardware [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 584.088105] env[61987]: DEBUG nova.virt.hardware [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 584.088105] env[61987]: DEBUG nova.virt.hardware [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 584.088703] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f00b201-8e08-4110-a6a3-d6964e0e75eb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.100528] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b6aa4d3-1d56-4678-8d5b-8ba3f0f5961d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.132838] env[61987]: DEBUG nova.scheduler.client.report [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 584.448775] env[61987]: DEBUG oslo_concurrency.lockutils [None req-efff1759-dc77-4623-a6b0-5b38908831bd tempest-ListServersNegativeTestJSON-1893919141 tempest-ListServersNegativeTestJSON-1893919141-project-member] Lock "c9ce2911-8fc4-4aa8-883e-487ee06670ba" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.864s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.490019] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Releasing lock "refresh_cache-7a94f776-579d-41e8-94aa-6095b3fd69c6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.490019] env[61987]: DEBUG nova.compute.manager [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 584.490019] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 584.490019] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7193b41c-73b8-44c5-85d0-f35e57c121a4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.499132] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d8e09f6-67f2-40cd-8427-2fed4a6a69a0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.526182] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7a94f776-579d-41e8-94aa-6095b3fd69c6 could not be found. [ 584.526182] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 584.526182] env[61987]: INFO nova.compute.manager [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Took 0.03 seconds to destroy the instance on the hypervisor. [ 584.526182] env[61987]: DEBUG oslo.service.loopingcall [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 584.526182] env[61987]: DEBUG nova.compute.manager [-] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 584.526182] env[61987]: DEBUG nova.network.neutron [-] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 584.545930] env[61987]: DEBUG nova.network.neutron [-] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 584.639222] env[61987]: DEBUG oslo_concurrency.lockutils [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.637s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.639571] env[61987]: DEBUG nova.compute.manager [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 584.644573] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.942s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.644573] env[61987]: INFO nova.compute.claims [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 584.763357] env[61987]: ERROR nova.compute.manager [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1aae5742-c738-447d-b60d-05f7c8f67871, please check neutron logs for more information. [ 584.763357] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 584.763357] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 584.763357] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 584.763357] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 584.763357] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 584.763357] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 584.763357] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 584.763357] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.763357] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 584.763357] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.763357] env[61987]: ERROR nova.compute.manager raise self.value [ 584.763357] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 584.763357] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 584.763357] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.763357] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 584.764027] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.764027] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 584.764027] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1aae5742-c738-447d-b60d-05f7c8f67871, please check neutron logs for more information. [ 584.764027] env[61987]: ERROR nova.compute.manager [ 584.764027] env[61987]: Traceback (most recent call last): [ 584.764027] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 584.764027] env[61987]: listener.cb(fileno) [ 584.764027] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.764027] env[61987]: result = function(*args, **kwargs) [ 584.764027] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 584.764027] env[61987]: return func(*args, **kwargs) [ 584.764027] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 584.764027] env[61987]: raise e [ 584.764027] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 584.764027] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 584.764027] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 584.764027] env[61987]: created_port_ids = self._update_ports_for_instance( [ 584.764027] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 584.764027] env[61987]: with excutils.save_and_reraise_exception(): [ 584.764027] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.764027] env[61987]: self.force_reraise() [ 584.764027] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.764027] env[61987]: raise self.value [ 584.764027] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 584.764027] env[61987]: updated_port = self._update_port( [ 584.764027] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.764027] env[61987]: _ensure_no_port_binding_failure(port) [ 584.764027] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.764027] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 584.764975] env[61987]: nova.exception.PortBindingFailed: Binding failed for port 1aae5742-c738-447d-b60d-05f7c8f67871, please check neutron logs for more information. [ 584.764975] env[61987]: Removing descriptor: 18 [ 584.764975] env[61987]: ERROR nova.compute.manager [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1aae5742-c738-447d-b60d-05f7c8f67871, please check neutron logs for more information. [ 584.764975] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Traceback (most recent call last): [ 584.764975] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 584.764975] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] yield resources [ 584.764975] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 584.764975] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] self.driver.spawn(context, instance, image_meta, [ 584.764975] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 584.764975] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] self._vmops.spawn(context, instance, image_meta, injected_files, [ 584.764975] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 584.764975] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] vm_ref = self.build_virtual_machine(instance, [ 584.765409] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 584.765409] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] vif_infos = vmwarevif.get_vif_info(self._session, [ 584.765409] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 584.765409] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] for vif in network_info: [ 584.765409] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 584.765409] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] return self._sync_wrapper(fn, *args, **kwargs) [ 584.765409] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 584.765409] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] self.wait() [ 584.765409] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 584.765409] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] self[:] = self._gt.wait() [ 584.765409] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 584.765409] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] return self._exit_event.wait() [ 584.765409] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 584.765853] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] result = hub.switch() [ 584.765853] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 584.765853] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] return self.greenlet.switch() [ 584.765853] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.765853] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] result = function(*args, **kwargs) [ 584.765853] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 584.765853] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] return func(*args, **kwargs) [ 584.765853] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 584.765853] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] raise e [ 584.765853] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 584.765853] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] nwinfo = self.network_api.allocate_for_instance( [ 584.765853] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 584.765853] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] created_port_ids = self._update_ports_for_instance( [ 584.766255] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 584.766255] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] with excutils.save_and_reraise_exception(): [ 584.766255] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.766255] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] self.force_reraise() [ 584.766255] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.766255] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] raise self.value [ 584.766255] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 584.766255] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] updated_port = self._update_port( [ 584.766255] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.766255] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] _ensure_no_port_binding_failure(port) [ 584.766255] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.766255] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] raise exception.PortBindingFailed(port_id=port['id']) [ 584.766633] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] nova.exception.PortBindingFailed: Binding failed for port 1aae5742-c738-447d-b60d-05f7c8f67871, please check neutron logs for more information. [ 584.766633] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] [ 584.766633] env[61987]: INFO nova.compute.manager [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Terminating instance [ 585.047093] env[61987]: DEBUG nova.network.neutron [-] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.148633] env[61987]: DEBUG nova.compute.utils [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 585.151967] env[61987]: DEBUG nova.compute.manager [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Not allocating networking since 'none' was specified. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 585.268966] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Acquiring lock "refresh_cache-938dab0c-9b30-404b-bca2-a1280a252584" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.269162] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Acquired lock "refresh_cache-938dab0c-9b30-404b-bca2-a1280a252584" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.269337] env[61987]: DEBUG nova.network.neutron [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 585.552021] env[61987]: INFO nova.compute.manager [-] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Took 1.03 seconds to deallocate network for instance. [ 585.554527] env[61987]: DEBUG nova.compute.claims [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 585.554698] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.652689] env[61987]: DEBUG nova.compute.manager [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 585.791861] env[61987]: DEBUG nova.network.neutron [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 585.891333] env[61987]: DEBUG nova.network.neutron [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.109613] env[61987]: DEBUG nova.compute.manager [req-fe5c6c12-5328-4295-b713-79ffec1bfc12 req-4f008dd5-311c-4918-bbb3-546e538d1356 service nova] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Received event network-changed-1aae5742-c738-447d-b60d-05f7c8f67871 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 586.109805] env[61987]: DEBUG nova.compute.manager [req-fe5c6c12-5328-4295-b713-79ffec1bfc12 req-4f008dd5-311c-4918-bbb3-546e538d1356 service nova] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Refreshing instance network info cache due to event network-changed-1aae5742-c738-447d-b60d-05f7c8f67871. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 586.109995] env[61987]: DEBUG oslo_concurrency.lockutils [req-fe5c6c12-5328-4295-b713-79ffec1bfc12 req-4f008dd5-311c-4918-bbb3-546e538d1356 service nova] Acquiring lock "refresh_cache-938dab0c-9b30-404b-bca2-a1280a252584" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.122058] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce00a87-b557-4c1f-a48d-0eaac819eca8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.130498] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea9e7c06-ce56-4aba-aa25-71d1a67fc471 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.173230] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5381ded0-b665-49f8-87d7-31d456709246 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.178329] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43b34597-4229-459c-baa0-52a60a69fec9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.192480] env[61987]: DEBUG nova.compute.provider_tree [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 586.398239] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Releasing lock "refresh_cache-938dab0c-9b30-404b-bca2-a1280a252584" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.399028] env[61987]: DEBUG nova.compute.manager [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 586.399028] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 586.400907] env[61987]: DEBUG oslo_concurrency.lockutils [req-fe5c6c12-5328-4295-b713-79ffec1bfc12 req-4f008dd5-311c-4918-bbb3-546e538d1356 service nova] Acquired lock "refresh_cache-938dab0c-9b30-404b-bca2-a1280a252584" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.401109] env[61987]: DEBUG nova.network.neutron [req-fe5c6c12-5328-4295-b713-79ffec1bfc12 req-4f008dd5-311c-4918-bbb3-546e538d1356 service nova] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Refreshing network info cache for port 1aae5742-c738-447d-b60d-05f7c8f67871 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 586.404297] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f59a16e4-ec13-4f1d-aeea-218da6bc5e57 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.424046] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5eb0cfb-96a8-4804-8ac1-a25101d69168 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.448375] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 938dab0c-9b30-404b-bca2-a1280a252584 could not be found. [ 586.448631] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 586.448895] env[61987]: INFO nova.compute.manager [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Took 0.05 seconds to destroy the instance on the hypervisor. [ 586.449161] env[61987]: DEBUG oslo.service.loopingcall [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 586.449161] env[61987]: DEBUG nova.compute.manager [-] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 586.449357] env[61987]: DEBUG nova.network.neutron [-] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 586.475020] env[61987]: DEBUG nova.network.neutron [-] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 586.671457] env[61987]: DEBUG nova.compute.manager [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 586.693017] env[61987]: DEBUG nova.virt.hardware [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 586.693387] env[61987]: DEBUG nova.virt.hardware [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 586.693561] env[61987]: DEBUG nova.virt.hardware [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 586.693746] env[61987]: DEBUG nova.virt.hardware [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 586.693893] env[61987]: DEBUG nova.virt.hardware [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 586.694052] env[61987]: DEBUG nova.virt.hardware [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 586.694256] env[61987]: DEBUG nova.virt.hardware [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 586.694411] env[61987]: DEBUG nova.virt.hardware [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 586.694615] env[61987]: DEBUG nova.virt.hardware [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 586.694788] env[61987]: DEBUG nova.virt.hardware [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 586.694960] env[61987]: DEBUG nova.virt.hardware [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 586.695779] env[61987]: DEBUG nova.scheduler.client.report [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 586.699697] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a5ce071-b871-4b2d-a1cc-9591975ea334 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.707546] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dd70377-1f73-49a6-8374-b0ec574dae38 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.726017] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Instance VIF info [] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 586.735110] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 586.735987] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-78685230-3a24-428f-a095-1621a70e8073 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.749234] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Created folder: OpenStack in parent group-v4. [ 586.749415] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Creating folder: Project (47e493e468d44bdabf77d61f12418c0f). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 586.749815] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-37b01c63-66e4-44c0-b48f-48aa60419c54 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.761044] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Created folder: Project (47e493e468d44bdabf77d61f12418c0f) in parent group-v234219. [ 586.761259] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Creating folder: Instances. Parent ref: group-v234220. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 586.761507] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-04949675-0a00-485c-b0eb-0ad004cb933d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.770221] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Created folder: Instances in parent group-v234220. [ 586.770455] env[61987]: DEBUG oslo.service.loopingcall [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 586.770695] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 586.770878] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aa02f5a0-0943-4ae0-b0cf-70af165e1960 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.789767] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 586.789767] env[61987]: value = "task-1061233" [ 586.789767] env[61987]: _type = "Task" [ 586.789767] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.798246] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061233, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.936399] env[61987]: DEBUG nova.network.neutron [req-fe5c6c12-5328-4295-b713-79ffec1bfc12 req-4f008dd5-311c-4918-bbb3-546e538d1356 service nova] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 586.978786] env[61987]: DEBUG nova.network.neutron [-] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.016311] env[61987]: DEBUG nova.network.neutron [req-fe5c6c12-5328-4295-b713-79ffec1bfc12 req-4f008dd5-311c-4918-bbb3-546e538d1356 service nova] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.204074] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.562s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.204569] env[61987]: DEBUG nova.compute.manager [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 587.207365] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.747s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.208895] env[61987]: INFO nova.compute.claims [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 587.299698] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061233, 'name': CreateVM_Task, 'duration_secs': 0.289112} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.299912] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 587.300914] env[61987]: DEBUG oslo_vmware.service [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5511e3f0-f7f0-479d-b01f-e2fae9f3ee3f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.307961] env[61987]: DEBUG oslo_concurrency.lockutils [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.307961] env[61987]: DEBUG oslo_concurrency.lockutils [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.307961] env[61987]: DEBUG oslo_concurrency.lockutils [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 587.308407] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-201c2fcd-6f18-43d8-98eb-9a93808d0833 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.312916] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Waiting for the task: (returnval){ [ 587.312916] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5218b5e4-eefd-24ec-a373-45b48a32dd4f" [ 587.312916] env[61987]: _type = "Task" [ 587.312916] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.321812] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5218b5e4-eefd-24ec-a373-45b48a32dd4f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.358293] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Acquiring lock "e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.358667] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Lock "e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.483249] env[61987]: INFO nova.compute.manager [-] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Took 1.03 seconds to deallocate network for instance. [ 587.486110] env[61987]: DEBUG nova.compute.claims [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 587.486218] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.521581] env[61987]: DEBUG oslo_concurrency.lockutils [req-fe5c6c12-5328-4295-b713-79ffec1bfc12 req-4f008dd5-311c-4918-bbb3-546e538d1356 service nova] Releasing lock "refresh_cache-938dab0c-9b30-404b-bca2-a1280a252584" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.521847] env[61987]: DEBUG nova.compute.manager [req-fe5c6c12-5328-4295-b713-79ffec1bfc12 req-4f008dd5-311c-4918-bbb3-546e538d1356 service nova] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Received event network-vif-deleted-1aae5742-c738-447d-b60d-05f7c8f67871 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 587.716349] env[61987]: DEBUG nova.compute.utils [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 587.717722] env[61987]: DEBUG nova.compute.manager [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 587.717885] env[61987]: DEBUG nova.network.neutron [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 587.768066] env[61987]: DEBUG nova.policy [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'abf116080a96445894baa4857597201d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cef456f2a7784b2c88c4ccb07db9005b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 587.824655] env[61987]: DEBUG oslo_concurrency.lockutils [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.824927] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 587.825186] env[61987]: DEBUG oslo_concurrency.lockutils [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.825366] env[61987]: DEBUG oslo_concurrency.lockutils [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.825769] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 587.826013] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a29c338c-03da-438b-ba26-657a52a1155f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.843132] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 587.843320] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 587.844096] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b3b4623-cf77-4666-8173-8839119b15ab {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.850407] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-589972ce-79f2-4e94-bf27-d4e2b888d196 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.856208] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Waiting for the task: (returnval){ [ 587.856208] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5270f959-9680-2ec5-902d-286d982ee1d0" [ 587.856208] env[61987]: _type = "Task" [ 587.856208] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.864806] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5270f959-9680-2ec5-902d-286d982ee1d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 588.138074] env[61987]: DEBUG nova.network.neutron [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Successfully created port: f516ce51-3f8d-4bf0-a708-95b2c95837f4 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 588.221897] env[61987]: DEBUG nova.compute.manager [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 588.371811] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Preparing fetch location {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 588.371811] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Creating directory with path [datastore2] vmware_temp/22c05e7d-4cff-4110-bc3f-2273ff450858/7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 588.371811] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9053143c-5bc7-4dd7-8741-fc916fae20c2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.397660] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Created directory with path [datastore2] vmware_temp/22c05e7d-4cff-4110-bc3f-2273ff450858/7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 588.397660] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Fetch image to [datastore2] vmware_temp/22c05e7d-4cff-4110-bc3f-2273ff450858/7bf8969f-d179-42af-9e15-527d5449f929/tmp-sparse.vmdk {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 588.397660] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Downloading image file data 7bf8969f-d179-42af-9e15-527d5449f929 to [datastore2] vmware_temp/22c05e7d-4cff-4110-bc3f-2273ff450858/7bf8969f-d179-42af-9e15-527d5449f929/tmp-sparse.vmdk on the data store datastore2 {{(pid=61987) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 588.397660] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be2539f7-b40f-4592-9f42-12e3f8971080 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.412319] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f3f3d44-d026-45af-90c6-90bb4b648ec8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.426153] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd0a33b3-a6fe-453e-8b59-9228f7134dc4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.463228] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60ba2401-daee-4a48-bcf6-327ac8c05de2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.470037] env[61987]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-e2f60917-1648-467a-9ef0-282d20730e7e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.493444] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Downloading image file data 7bf8969f-d179-42af-9e15-527d5449f929 to the data store datastore2 {{(pid=61987) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 588.566621] env[61987]: DEBUG oslo_vmware.rw_handles [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/22c05e7d-4cff-4110-bc3f-2273ff450858/7bf8969f-d179-42af-9e15-527d5449f929/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61987) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 588.825360] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90ba7b58-7442-4f6c-8ca4-42eaeeb523de {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.834786] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3696e4e3-6fba-407b-a9d5-69fb9c765405 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.874766] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3bd5dae-3555-4290-87fd-04316d05bfca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.883756] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efaa4c98-fab0-4562-8ad6-d9556fe3b283 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.899762] env[61987]: DEBUG nova.compute.provider_tree [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 589.017304] env[61987]: DEBUG nova.compute.manager [req-12f85f9a-84c1-4a29-b561-0246aa300f64 req-26b8571f-e8d9-40ce-96ce-d8bbd3bd0d93 service nova] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Received event network-changed-f516ce51-3f8d-4bf0-a708-95b2c95837f4 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 589.017512] env[61987]: DEBUG nova.compute.manager [req-12f85f9a-84c1-4a29-b561-0246aa300f64 req-26b8571f-e8d9-40ce-96ce-d8bbd3bd0d93 service nova] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Refreshing instance network info cache due to event network-changed-f516ce51-3f8d-4bf0-a708-95b2c95837f4. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 589.017761] env[61987]: DEBUG oslo_concurrency.lockutils [req-12f85f9a-84c1-4a29-b561-0246aa300f64 req-26b8571f-e8d9-40ce-96ce-d8bbd3bd0d93 service nova] Acquiring lock "refresh_cache-7961bd0d-de24-4e13-94a3-79c48c593813" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.018023] env[61987]: DEBUG oslo_concurrency.lockutils [req-12f85f9a-84c1-4a29-b561-0246aa300f64 req-26b8571f-e8d9-40ce-96ce-d8bbd3bd0d93 service nova] Acquired lock "refresh_cache-7961bd0d-de24-4e13-94a3-79c48c593813" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.018145] env[61987]: DEBUG nova.network.neutron [req-12f85f9a-84c1-4a29-b561-0246aa300f64 req-26b8571f-e8d9-40ce-96ce-d8bbd3bd0d93 service nova] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Refreshing network info cache for port f516ce51-3f8d-4bf0-a708-95b2c95837f4 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 589.241638] env[61987]: DEBUG nova.compute.manager [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 589.260638] env[61987]: ERROR nova.compute.manager [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f516ce51-3f8d-4bf0-a708-95b2c95837f4, please check neutron logs for more information. [ 589.260638] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 589.260638] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 589.260638] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 589.260638] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 589.260638] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 589.260638] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 589.260638] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 589.260638] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.260638] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 589.260638] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.260638] env[61987]: ERROR nova.compute.manager raise self.value [ 589.260638] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 589.260638] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 589.260638] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.260638] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 589.262150] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.262150] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 589.262150] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f516ce51-3f8d-4bf0-a708-95b2c95837f4, please check neutron logs for more information. [ 589.262150] env[61987]: ERROR nova.compute.manager [ 589.262150] env[61987]: Traceback (most recent call last): [ 589.262150] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 589.262150] env[61987]: listener.cb(fileno) [ 589.262150] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.262150] env[61987]: result = function(*args, **kwargs) [ 589.262150] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 589.262150] env[61987]: return func(*args, **kwargs) [ 589.262150] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 589.262150] env[61987]: raise e [ 589.262150] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 589.262150] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 589.262150] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 589.262150] env[61987]: created_port_ids = self._update_ports_for_instance( [ 589.262150] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 589.262150] env[61987]: with excutils.save_and_reraise_exception(): [ 589.262150] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.262150] env[61987]: self.force_reraise() [ 589.262150] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.262150] env[61987]: raise self.value [ 589.262150] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 589.262150] env[61987]: updated_port = self._update_port( [ 589.262150] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.262150] env[61987]: _ensure_no_port_binding_failure(port) [ 589.262150] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.262150] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 589.263811] env[61987]: nova.exception.PortBindingFailed: Binding failed for port f516ce51-3f8d-4bf0-a708-95b2c95837f4, please check neutron logs for more information. [ 589.263811] env[61987]: Removing descriptor: 18 [ 589.267193] env[61987]: DEBUG nova.virt.hardware [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 589.267499] env[61987]: DEBUG nova.virt.hardware [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 589.267695] env[61987]: DEBUG nova.virt.hardware [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 589.267920] env[61987]: DEBUG nova.virt.hardware [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 589.268250] env[61987]: DEBUG nova.virt.hardware [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 589.268320] env[61987]: DEBUG nova.virt.hardware [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 589.268530] env[61987]: DEBUG nova.virt.hardware [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 589.268723] env[61987]: DEBUG nova.virt.hardware [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 589.268916] env[61987]: DEBUG nova.virt.hardware [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 589.269125] env[61987]: DEBUG nova.virt.hardware [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 589.269353] env[61987]: DEBUG nova.virt.hardware [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 589.270295] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-151e1365-69de-4044-923f-580468f7096b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.279186] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0a68eb7-e1df-4a2f-a627-50d6da22411a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.299195] env[61987]: ERROR nova.compute.manager [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f516ce51-3f8d-4bf0-a708-95b2c95837f4, please check neutron logs for more information. [ 589.299195] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Traceback (most recent call last): [ 589.299195] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 589.299195] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] yield resources [ 589.299195] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 589.299195] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] self.driver.spawn(context, instance, image_meta, [ 589.299195] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 589.299195] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] self._vmops.spawn(context, instance, image_meta, injected_files, [ 589.299195] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 589.299195] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] vm_ref = self.build_virtual_machine(instance, [ 589.299195] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 589.300289] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] vif_infos = vmwarevif.get_vif_info(self._session, [ 589.300289] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 589.300289] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] for vif in network_info: [ 589.300289] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 589.300289] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] return self._sync_wrapper(fn, *args, **kwargs) [ 589.300289] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 589.300289] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] self.wait() [ 589.300289] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 589.300289] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] self[:] = self._gt.wait() [ 589.300289] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 589.300289] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] return self._exit_event.wait() [ 589.300289] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 589.300289] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] current.throw(*self._exc) [ 589.300812] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.300812] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] result = function(*args, **kwargs) [ 589.300812] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 589.300812] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] return func(*args, **kwargs) [ 589.300812] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 589.300812] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] raise e [ 589.300812] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 589.300812] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] nwinfo = self.network_api.allocate_for_instance( [ 589.300812] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 589.300812] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] created_port_ids = self._update_ports_for_instance( [ 589.300812] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 589.300812] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] with excutils.save_and_reraise_exception(): [ 589.300812] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.301306] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] self.force_reraise() [ 589.301306] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.301306] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] raise self.value [ 589.301306] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 589.301306] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] updated_port = self._update_port( [ 589.301306] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.301306] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] _ensure_no_port_binding_failure(port) [ 589.301306] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.301306] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] raise exception.PortBindingFailed(port_id=port['id']) [ 589.301306] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] nova.exception.PortBindingFailed: Binding failed for port f516ce51-3f8d-4bf0-a708-95b2c95837f4, please check neutron logs for more information. [ 589.301306] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] [ 589.301306] env[61987]: INFO nova.compute.manager [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Terminating instance [ 589.357056] env[61987]: DEBUG oslo_vmware.rw_handles [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Completed reading data from the image iterator. {{(pid=61987) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 589.357056] env[61987]: DEBUG oslo_vmware.rw_handles [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/22c05e7d-4cff-4110-bc3f-2273ff450858/7bf8969f-d179-42af-9e15-527d5449f929/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61987) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 589.404793] env[61987]: DEBUG nova.scheduler.client.report [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 589.469136] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Downloaded image file data 7bf8969f-d179-42af-9e15-527d5449f929 to vmware_temp/22c05e7d-4cff-4110-bc3f-2273ff450858/7bf8969f-d179-42af-9e15-527d5449f929/tmp-sparse.vmdk on the data store datastore2 {{(pid=61987) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 589.471424] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Caching image {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 589.471709] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Copying Virtual Disk [datastore2] vmware_temp/22c05e7d-4cff-4110-bc3f-2273ff450858/7bf8969f-d179-42af-9e15-527d5449f929/tmp-sparse.vmdk to [datastore2] vmware_temp/22c05e7d-4cff-4110-bc3f-2273ff450858/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 589.472012] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-10ab1e26-6232-4c23-b3d4-da64f9f5b642 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.481076] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Waiting for the task: (returnval){ [ 589.481076] env[61987]: value = "task-1061234" [ 589.481076] env[61987]: _type = "Task" [ 589.481076] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 589.489707] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061234, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 589.540415] env[61987]: DEBUG nova.network.neutron [req-12f85f9a-84c1-4a29-b561-0246aa300f64 req-26b8571f-e8d9-40ce-96ce-d8bbd3bd0d93 service nova] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 589.633107] env[61987]: DEBUG nova.network.neutron [req-12f85f9a-84c1-4a29-b561-0246aa300f64 req-26b8571f-e8d9-40ce-96ce-d8bbd3bd0d93 service nova] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.812362] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Acquiring lock "refresh_cache-7961bd0d-de24-4e13-94a3-79c48c593813" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.913152] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.706s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.913681] env[61987]: DEBUG nova.compute.manager [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 589.916400] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.490s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.917780] env[61987]: INFO nova.compute.claims [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 589.992777] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061234, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 590.140277] env[61987]: DEBUG oslo_concurrency.lockutils [req-12f85f9a-84c1-4a29-b561-0246aa300f64 req-26b8571f-e8d9-40ce-96ce-d8bbd3bd0d93 service nova] Releasing lock "refresh_cache-7961bd0d-de24-4e13-94a3-79c48c593813" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.140775] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Acquired lock "refresh_cache-7961bd0d-de24-4e13-94a3-79c48c593813" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.141011] env[61987]: DEBUG nova.network.neutron [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 590.421839] env[61987]: DEBUG nova.compute.utils [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 590.425120] env[61987]: DEBUG nova.compute.manager [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 590.425295] env[61987]: DEBUG nova.network.neutron [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 590.471034] env[61987]: DEBUG nova.policy [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '351d1ff9df734b828914e25c89a2cade', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bde0bf5c09724823ab56ac4a899a3561', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 590.494238] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061234, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.795779} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 590.494507] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Copied Virtual Disk [datastore2] vmware_temp/22c05e7d-4cff-4110-bc3f-2273ff450858/7bf8969f-d179-42af-9e15-527d5449f929/tmp-sparse.vmdk to [datastore2] vmware_temp/22c05e7d-4cff-4110-bc3f-2273ff450858/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 590.494745] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Deleting the datastore file [datastore2] vmware_temp/22c05e7d-4cff-4110-bc3f-2273ff450858/7bf8969f-d179-42af-9e15-527d5449f929/tmp-sparse.vmdk {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 590.494922] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c91550ce-ec7f-4584-a2df-e9b9005b24d0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.502667] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Waiting for the task: (returnval){ [ 590.502667] env[61987]: value = "task-1061235" [ 590.502667] env[61987]: _type = "Task" [ 590.502667] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 590.510979] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061235, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 590.666708] env[61987]: DEBUG nova.network.neutron [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 590.786314] env[61987]: DEBUG nova.network.neutron [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.928068] env[61987]: DEBUG nova.compute.manager [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 591.015083] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061235, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.022962} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 591.016161] env[61987]: DEBUG nova.network.neutron [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Successfully created port: e99cf30c-90d8-4653-9d2c-4abd600020bd {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 591.017658] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 591.017871] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Moving file from [datastore2] vmware_temp/22c05e7d-4cff-4110-bc3f-2273ff450858/7bf8969f-d179-42af-9e15-527d5449f929 to [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929. {{(pid=61987) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 591.021308] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-6a134308-7232-4462-af4a-62b18b4d3554 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.030055] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Waiting for the task: (returnval){ [ 591.030055] env[61987]: value = "task-1061236" [ 591.030055] env[61987]: _type = "Task" [ 591.030055] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 591.041239] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061236, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.066082] env[61987]: DEBUG nova.compute.manager [req-2f872c49-64e6-40ab-98f1-4664394f692d req-e9c9d4b9-2110-46dd-a09c-eea901f28a46 service nova] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Received event network-vif-deleted-f516ce51-3f8d-4bf0-a708-95b2c95837f4 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 591.288481] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Releasing lock "refresh_cache-7961bd0d-de24-4e13-94a3-79c48c593813" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.288824] env[61987]: DEBUG nova.compute.manager [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 591.289028] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 591.289342] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8627e395-72a9-4eec-9673-5062b93b38ce {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.301727] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-966d3d2d-2515-4b0e-8c6a-fdae70f2538a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.332589] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7961bd0d-de24-4e13-94a3-79c48c593813 could not be found. [ 591.333705] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 591.333705] env[61987]: INFO nova.compute.manager [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Took 0.04 seconds to destroy the instance on the hypervisor. [ 591.333705] env[61987]: DEBUG oslo.service.loopingcall [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 591.336603] env[61987]: DEBUG nova.compute.manager [-] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 591.336603] env[61987]: DEBUG nova.network.neutron [-] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 591.361143] env[61987]: DEBUG nova.network.neutron [-] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 591.431156] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d58c57fe-8454-4d2f-bb5a-384c3e911045 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.443838] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-580a0736-5913-4eb0-9b4a-ef1c4de4e65a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.479424] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32edf0ca-d1b5-4415-ac18-ca46f21144ec {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.487688] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c334106-ea20-459d-95c8-594befe60662 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.502204] env[61987]: DEBUG nova.compute.provider_tree [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.539160] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061236, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.022684} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 591.539641] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] File moved {{(pid=61987) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 591.539992] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Cleaning up location [datastore2] vmware_temp/22c05e7d-4cff-4110-bc3f-2273ff450858 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 591.540315] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Deleting the datastore file [datastore2] vmware_temp/22c05e7d-4cff-4110-bc3f-2273ff450858 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 591.540706] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-276f2a1c-9088-48fa-946f-5e0496661ef9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.549927] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Waiting for the task: (returnval){ [ 591.549927] env[61987]: value = "task-1061237" [ 591.549927] env[61987]: _type = "Task" [ 591.549927] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 591.556709] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061237, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 591.864077] env[61987]: DEBUG nova.network.neutron [-] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.942081] env[61987]: DEBUG nova.compute.manager [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 591.986032] env[61987]: DEBUG nova.virt.hardware [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 591.986032] env[61987]: DEBUG nova.virt.hardware [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 591.986032] env[61987]: DEBUG nova.virt.hardware [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 591.986216] env[61987]: DEBUG nova.virt.hardware [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 591.986216] env[61987]: DEBUG nova.virt.hardware [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 591.986216] env[61987]: DEBUG nova.virt.hardware [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 591.986216] env[61987]: DEBUG nova.virt.hardware [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 591.986216] env[61987]: DEBUG nova.virt.hardware [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 591.990018] env[61987]: DEBUG nova.virt.hardware [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 591.990018] env[61987]: DEBUG nova.virt.hardware [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 591.990018] env[61987]: DEBUG nova.virt.hardware [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 591.990018] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e6a3af2-aac0-4129-a217-752781437e70 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.996700] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d3ba77d-4768-4a90-90d2-66e7ac6ad2fd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.010886] env[61987]: DEBUG nova.scheduler.client.report [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 592.061029] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061237, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.022974} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 592.061540] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 592.062310] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6fe615b8-6ff1-4128-8a33-7e6b28c3cc3d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.072043] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Waiting for the task: (returnval){ [ 592.072043] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]522bb6ea-76a5-120b-5f82-1160dfbcbced" [ 592.072043] env[61987]: _type = "Task" [ 592.072043] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.079645] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]522bb6ea-76a5-120b-5f82-1160dfbcbced, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.204750] env[61987]: ERROR nova.compute.manager [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e99cf30c-90d8-4653-9d2c-4abd600020bd, please check neutron logs for more information. [ 592.204750] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 592.204750] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 592.204750] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 592.204750] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 592.204750] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 592.204750] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 592.204750] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 592.204750] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.204750] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 592.204750] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.204750] env[61987]: ERROR nova.compute.manager raise self.value [ 592.204750] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 592.204750] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 592.204750] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.204750] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 592.205325] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.205325] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 592.205325] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e99cf30c-90d8-4653-9d2c-4abd600020bd, please check neutron logs for more information. [ 592.205325] env[61987]: ERROR nova.compute.manager [ 592.205325] env[61987]: Traceback (most recent call last): [ 592.205325] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 592.205325] env[61987]: listener.cb(fileno) [ 592.205325] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.205325] env[61987]: result = function(*args, **kwargs) [ 592.205325] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 592.205325] env[61987]: return func(*args, **kwargs) [ 592.205325] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 592.205325] env[61987]: raise e [ 592.205325] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 592.205325] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 592.205325] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 592.205325] env[61987]: created_port_ids = self._update_ports_for_instance( [ 592.205325] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 592.205325] env[61987]: with excutils.save_and_reraise_exception(): [ 592.205325] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.205325] env[61987]: self.force_reraise() [ 592.205325] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.205325] env[61987]: raise self.value [ 592.205325] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 592.205325] env[61987]: updated_port = self._update_port( [ 592.205325] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.205325] env[61987]: _ensure_no_port_binding_failure(port) [ 592.205325] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.205325] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 592.206251] env[61987]: nova.exception.PortBindingFailed: Binding failed for port e99cf30c-90d8-4653-9d2c-4abd600020bd, please check neutron logs for more information. [ 592.206251] env[61987]: Removing descriptor: 18 [ 592.206251] env[61987]: ERROR nova.compute.manager [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e99cf30c-90d8-4653-9d2c-4abd600020bd, please check neutron logs for more information. [ 592.206251] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Traceback (most recent call last): [ 592.206251] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 592.206251] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] yield resources [ 592.206251] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 592.206251] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] self.driver.spawn(context, instance, image_meta, [ 592.206251] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 592.206251] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 592.206251] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 592.206251] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] vm_ref = self.build_virtual_machine(instance, [ 592.206669] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 592.206669] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] vif_infos = vmwarevif.get_vif_info(self._session, [ 592.206669] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 592.206669] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] for vif in network_info: [ 592.206669] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 592.206669] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] return self._sync_wrapper(fn, *args, **kwargs) [ 592.206669] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 592.206669] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] self.wait() [ 592.206669] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 592.206669] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] self[:] = self._gt.wait() [ 592.206669] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 592.206669] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] return self._exit_event.wait() [ 592.206669] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 592.207123] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] result = hub.switch() [ 592.207123] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 592.207123] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] return self.greenlet.switch() [ 592.207123] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.207123] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] result = function(*args, **kwargs) [ 592.207123] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 592.207123] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] return func(*args, **kwargs) [ 592.207123] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 592.207123] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] raise e [ 592.207123] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 592.207123] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] nwinfo = self.network_api.allocate_for_instance( [ 592.207123] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 592.207123] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] created_port_ids = self._update_ports_for_instance( [ 592.207638] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 592.207638] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] with excutils.save_and_reraise_exception(): [ 592.207638] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.207638] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] self.force_reraise() [ 592.207638] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.207638] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] raise self.value [ 592.207638] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 592.207638] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] updated_port = self._update_port( [ 592.207638] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.207638] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] _ensure_no_port_binding_failure(port) [ 592.207638] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.207638] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] raise exception.PortBindingFailed(port_id=port['id']) [ 592.208025] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] nova.exception.PortBindingFailed: Binding failed for port e99cf30c-90d8-4653-9d2c-4abd600020bd, please check neutron logs for more information. [ 592.208025] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] [ 592.208025] env[61987]: INFO nova.compute.manager [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Terminating instance [ 592.366514] env[61987]: INFO nova.compute.manager [-] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Took 1.03 seconds to deallocate network for instance. [ 592.369082] env[61987]: DEBUG nova.compute.claims [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 592.369213] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.518915] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.602s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.519833] env[61987]: DEBUG nova.compute.manager [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 592.522454] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.076s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.523861] env[61987]: INFO nova.compute.claims [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 592.579922] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]522bb6ea-76a5-120b-5f82-1160dfbcbced, 'name': SearchDatastore_Task, 'duration_secs': 0.00936} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 592.580211] env[61987]: DEBUG oslo_concurrency.lockutils [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.580468] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 6bf0d1f1-d079-4956-bc9a-86ba66885f3a/6bf0d1f1-d079-4956-bc9a-86ba66885f3a.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 592.580714] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-095a3775-950e-4db4-9ae6-5dd1d3dc9277 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.588323] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Waiting for the task: (returnval){ [ 592.588323] env[61987]: value = "task-1061238" [ 592.588323] env[61987]: _type = "Task" [ 592.588323] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 592.595803] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061238, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 592.712997] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Acquiring lock "refresh_cache-bac5a3bc-5a58-4c1d-abb8-1aaf6151358a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.713248] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Acquired lock "refresh_cache-bac5a3bc-5a58-4c1d-abb8-1aaf6151358a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.713447] env[61987]: DEBUG nova.network.neutron [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 593.030306] env[61987]: DEBUG nova.compute.utils [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 593.032832] env[61987]: DEBUG nova.compute.manager [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 593.034247] env[61987]: DEBUG nova.network.neutron [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 593.089026] env[61987]: DEBUG nova.policy [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1dad4ca9c9f142aaa74c2791586499da', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2d401c1631e24de3be089251e94fab31', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 593.095504] env[61987]: DEBUG nova.compute.manager [req-5588c062-ca22-4553-a35e-d3a0fe827c4d req-7bfc1636-dc95-485d-a3fc-cf1ff29827bf service nova] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Received event network-changed-e99cf30c-90d8-4653-9d2c-4abd600020bd {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 593.095863] env[61987]: DEBUG nova.compute.manager [req-5588c062-ca22-4553-a35e-d3a0fe827c4d req-7bfc1636-dc95-485d-a3fc-cf1ff29827bf service nova] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Refreshing instance network info cache due to event network-changed-e99cf30c-90d8-4653-9d2c-4abd600020bd. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 593.096212] env[61987]: DEBUG oslo_concurrency.lockutils [req-5588c062-ca22-4553-a35e-d3a0fe827c4d req-7bfc1636-dc95-485d-a3fc-cf1ff29827bf service nova] Acquiring lock "refresh_cache-bac5a3bc-5a58-4c1d-abb8-1aaf6151358a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 593.103378] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061238, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.445796} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.103716] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 6bf0d1f1-d079-4956-bc9a-86ba66885f3a/6bf0d1f1-d079-4956-bc9a-86ba66885f3a.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 593.104463] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 593.104816] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-56961cbf-8585-45fd-a40a-299fdc6b1b84 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.111628] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Waiting for the task: (returnval){ [ 593.111628] env[61987]: value = "task-1061239" [ 593.111628] env[61987]: _type = "Task" [ 593.111628] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.120074] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061239, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.236796] env[61987]: DEBUG nova.network.neutron [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 593.403535] env[61987]: DEBUG nova.network.neutron [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.483853] env[61987]: DEBUG nova.network.neutron [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Successfully created port: 5f60a42f-4af4-440f-93a1-79dacf25140a {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 593.534044] env[61987]: DEBUG nova.compute.manager [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 593.625362] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061239, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068639} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 593.625617] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 593.626547] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07432449-f09e-4ddf-b793-dc0787daa3a2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.654093] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Reconfiguring VM instance instance-0000000f to attach disk [datastore2] 6bf0d1f1-d079-4956-bc9a-86ba66885f3a/6bf0d1f1-d079-4956-bc9a-86ba66885f3a.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 593.657147] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4fb7321c-ad8d-41f4-808d-5e5c9e029f63 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.679321] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Waiting for the task: (returnval){ [ 593.679321] env[61987]: value = "task-1061240" [ 593.679321] env[61987]: _type = "Task" [ 593.679321] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 593.691435] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061240, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 593.905741] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Releasing lock "refresh_cache-bac5a3bc-5a58-4c1d-abb8-1aaf6151358a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.906176] env[61987]: DEBUG nova.compute.manager [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 593.906364] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 593.906729] env[61987]: DEBUG oslo_concurrency.lockutils [req-5588c062-ca22-4553-a35e-d3a0fe827c4d req-7bfc1636-dc95-485d-a3fc-cf1ff29827bf service nova] Acquired lock "refresh_cache-bac5a3bc-5a58-4c1d-abb8-1aaf6151358a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.906903] env[61987]: DEBUG nova.network.neutron [req-5588c062-ca22-4553-a35e-d3a0fe827c4d req-7bfc1636-dc95-485d-a3fc-cf1ff29827bf service nova] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Refreshing network info cache for port e99cf30c-90d8-4653-9d2c-4abd600020bd {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 593.908014] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bfce51e0-da47-4b6c-9f95-61968e9b4d02 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.922374] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a53ce04-19f9-4e8b-a968-6bdef5955202 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.949150] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bac5a3bc-5a58-4c1d-abb8-1aaf6151358a could not be found. [ 593.949359] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 593.949536] env[61987]: INFO nova.compute.manager [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 593.949776] env[61987]: DEBUG oslo.service.loopingcall [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 593.949995] env[61987]: DEBUG nova.compute.manager [-] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 593.950107] env[61987]: DEBUG nova.network.neutron [-] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 593.970051] env[61987]: DEBUG nova.network.neutron [-] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.005974] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3adfcf83-d015-48a7-9882-1bb542277b44 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.014300] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bccd2dc-24cd-490d-8a98-8ceb961792f3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.051140] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc5e459-c616-4479-911a-c743e34a1f1f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.056941] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9491bc85-f841-46f6-bf0e-69547d007ff1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.071760] env[61987]: DEBUG nova.compute.provider_tree [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.189314] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061240, 'name': ReconfigVM_Task, 'duration_secs': 0.285404} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.189610] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Reconfigured VM instance instance-0000000f to attach disk [datastore2] 6bf0d1f1-d079-4956-bc9a-86ba66885f3a/6bf0d1f1-d079-4956-bc9a-86ba66885f3a.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 594.191074] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e249b489-ae9e-4a26-b259-444e76356af2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.197543] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Waiting for the task: (returnval){ [ 594.197543] env[61987]: value = "task-1061241" [ 594.197543] env[61987]: _type = "Task" [ 594.197543] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.206448] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061241, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.432615] env[61987]: DEBUG nova.network.neutron [req-5588c062-ca22-4553-a35e-d3a0fe827c4d req-7bfc1636-dc95-485d-a3fc-cf1ff29827bf service nova] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.471350] env[61987]: DEBUG nova.network.neutron [-] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.527396] env[61987]: ERROR nova.compute.manager [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5f60a42f-4af4-440f-93a1-79dacf25140a, please check neutron logs for more information. [ 594.527396] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 594.527396] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 594.527396] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 594.527396] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 594.527396] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 594.527396] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 594.527396] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 594.527396] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.527396] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 594.527396] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.527396] env[61987]: ERROR nova.compute.manager raise self.value [ 594.527396] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 594.527396] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 594.527396] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.527396] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 594.527950] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.527950] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 594.527950] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5f60a42f-4af4-440f-93a1-79dacf25140a, please check neutron logs for more information. [ 594.527950] env[61987]: ERROR nova.compute.manager [ 594.527950] env[61987]: Traceback (most recent call last): [ 594.527950] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 594.527950] env[61987]: listener.cb(fileno) [ 594.527950] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.527950] env[61987]: result = function(*args, **kwargs) [ 594.527950] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 594.527950] env[61987]: return func(*args, **kwargs) [ 594.527950] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 594.527950] env[61987]: raise e [ 594.527950] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 594.527950] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 594.527950] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 594.527950] env[61987]: created_port_ids = self._update_ports_for_instance( [ 594.527950] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 594.527950] env[61987]: with excutils.save_and_reraise_exception(): [ 594.527950] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.527950] env[61987]: self.force_reraise() [ 594.527950] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.527950] env[61987]: raise self.value [ 594.527950] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 594.527950] env[61987]: updated_port = self._update_port( [ 594.527950] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.527950] env[61987]: _ensure_no_port_binding_failure(port) [ 594.527950] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.527950] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 594.528897] env[61987]: nova.exception.PortBindingFailed: Binding failed for port 5f60a42f-4af4-440f-93a1-79dacf25140a, please check neutron logs for more information. [ 594.528897] env[61987]: Removing descriptor: 18 [ 594.548882] env[61987]: DEBUG nova.compute.manager [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 594.562592] env[61987]: DEBUG nova.network.neutron [req-5588c062-ca22-4553-a35e-d3a0fe827c4d req-7bfc1636-dc95-485d-a3fc-cf1ff29827bf service nova] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.573745] env[61987]: DEBUG nova.virt.hardware [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 594.574015] env[61987]: DEBUG nova.virt.hardware [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 594.574227] env[61987]: DEBUG nova.virt.hardware [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 594.574433] env[61987]: DEBUG nova.virt.hardware [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 594.574792] env[61987]: DEBUG nova.virt.hardware [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 594.574792] env[61987]: DEBUG nova.virt.hardware [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 594.574911] env[61987]: DEBUG nova.virt.hardware [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 594.575975] env[61987]: DEBUG nova.virt.hardware [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 594.575975] env[61987]: DEBUG nova.virt.hardware [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 594.575975] env[61987]: DEBUG nova.virt.hardware [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 594.575975] env[61987]: DEBUG nova.virt.hardware [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 594.576748] env[61987]: DEBUG nova.scheduler.client.report [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 594.580173] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-065dd7d5-5e15-41db-aba2-c1bd03b612b7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.590238] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ee2ce59-aefa-4a7a-9b25-f8f6560663a2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.603870] env[61987]: ERROR nova.compute.manager [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5f60a42f-4af4-440f-93a1-79dacf25140a, please check neutron logs for more information. [ 594.603870] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Traceback (most recent call last): [ 594.603870] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 594.603870] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] yield resources [ 594.603870] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 594.603870] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] self.driver.spawn(context, instance, image_meta, [ 594.603870] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 594.603870] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 594.603870] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 594.603870] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] vm_ref = self.build_virtual_machine(instance, [ 594.603870] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 594.604444] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] vif_infos = vmwarevif.get_vif_info(self._session, [ 594.604444] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 594.604444] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] for vif in network_info: [ 594.604444] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 594.604444] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] return self._sync_wrapper(fn, *args, **kwargs) [ 594.604444] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 594.604444] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] self.wait() [ 594.604444] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 594.604444] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] self[:] = self._gt.wait() [ 594.604444] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 594.604444] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] return self._exit_event.wait() [ 594.604444] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 594.604444] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] current.throw(*self._exc) [ 594.604993] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.604993] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] result = function(*args, **kwargs) [ 594.604993] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 594.604993] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] return func(*args, **kwargs) [ 594.604993] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 594.604993] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] raise e [ 594.604993] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 594.604993] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] nwinfo = self.network_api.allocate_for_instance( [ 594.604993] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 594.604993] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] created_port_ids = self._update_ports_for_instance( [ 594.604993] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 594.604993] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] with excutils.save_and_reraise_exception(): [ 594.604993] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.605466] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] self.force_reraise() [ 594.605466] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.605466] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] raise self.value [ 594.605466] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 594.605466] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] updated_port = self._update_port( [ 594.605466] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.605466] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] _ensure_no_port_binding_failure(port) [ 594.605466] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.605466] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] raise exception.PortBindingFailed(port_id=port['id']) [ 594.605466] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] nova.exception.PortBindingFailed: Binding failed for port 5f60a42f-4af4-440f-93a1-79dacf25140a, please check neutron logs for more information. [ 594.605466] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] [ 594.605466] env[61987]: INFO nova.compute.manager [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Terminating instance [ 594.708149] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061241, 'name': Rename_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.975479] env[61987]: INFO nova.compute.manager [-] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Took 1.02 seconds to deallocate network for instance. [ 594.976507] env[61987]: DEBUG nova.compute.claims [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 594.976912] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.065259] env[61987]: DEBUG oslo_concurrency.lockutils [req-5588c062-ca22-4553-a35e-d3a0fe827c4d req-7bfc1636-dc95-485d-a3fc-cf1ff29827bf service nova] Releasing lock "refresh_cache-bac5a3bc-5a58-4c1d-abb8-1aaf6151358a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.065595] env[61987]: DEBUG nova.compute.manager [req-5588c062-ca22-4553-a35e-d3a0fe827c4d req-7bfc1636-dc95-485d-a3fc-cf1ff29827bf service nova] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Received event network-vif-deleted-e99cf30c-90d8-4653-9d2c-4abd600020bd {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 595.083878] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.561s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.084390] env[61987]: DEBUG nova.compute.manager [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 595.086863] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.465s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.088232] env[61987]: INFO nova.compute.claims [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 595.108200] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Acquiring lock "refresh_cache-660e5e4c-0804-429b-ab18-9305a433391f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.108304] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Acquired lock "refresh_cache-660e5e4c-0804-429b-ab18-9305a433391f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.108443] env[61987]: DEBUG nova.network.neutron [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 595.130806] env[61987]: DEBUG nova.compute.manager [req-f9cf16e0-1e64-4127-b7fd-0d727c0a4e0a req-1b4147f8-6a54-4b00-8c44-f69a86adf161 service nova] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Received event network-changed-5f60a42f-4af4-440f-93a1-79dacf25140a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 595.130806] env[61987]: DEBUG nova.compute.manager [req-f9cf16e0-1e64-4127-b7fd-0d727c0a4e0a req-1b4147f8-6a54-4b00-8c44-f69a86adf161 service nova] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Refreshing instance network info cache due to event network-changed-5f60a42f-4af4-440f-93a1-79dacf25140a. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 595.130806] env[61987]: DEBUG oslo_concurrency.lockutils [req-f9cf16e0-1e64-4127-b7fd-0d727c0a4e0a req-1b4147f8-6a54-4b00-8c44-f69a86adf161 service nova] Acquiring lock "refresh_cache-660e5e4c-0804-429b-ab18-9305a433391f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.211588] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061241, 'name': Rename_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.592823] env[61987]: DEBUG nova.compute.utils [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 595.595610] env[61987]: DEBUG nova.compute.manager [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 595.595774] env[61987]: DEBUG nova.network.neutron [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 595.630351] env[61987]: DEBUG nova.network.neutron [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 595.684863] env[61987]: DEBUG nova.policy [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1dad4ca9c9f142aaa74c2791586499da', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2d401c1631e24de3be089251e94fab31', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 595.709553] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061241, 'name': Rename_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.742407] env[61987]: DEBUG nova.network.neutron [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.025671] env[61987]: DEBUG nova.network.neutron [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Successfully created port: 45aa1dda-6ba4-4927-9269-11fba215aa5a {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 596.097484] env[61987]: DEBUG nova.compute.manager [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 596.216018] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061241, 'name': Rename_Task, 'duration_secs': 1.877675} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.216018] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 596.216018] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2c2a4c6e-f6b2-4c21-984e-7e4a427758cb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.223805] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Waiting for the task: (returnval){ [ 596.223805] env[61987]: value = "task-1061242" [ 596.223805] env[61987]: _type = "Task" [ 596.223805] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.236199] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061242, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.248020] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Releasing lock "refresh_cache-660e5e4c-0804-429b-ab18-9305a433391f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.248020] env[61987]: DEBUG nova.compute.manager [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 596.248020] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 596.248020] env[61987]: DEBUG oslo_concurrency.lockutils [req-f9cf16e0-1e64-4127-b7fd-0d727c0a4e0a req-1b4147f8-6a54-4b00-8c44-f69a86adf161 service nova] Acquired lock "refresh_cache-660e5e4c-0804-429b-ab18-9305a433391f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.248020] env[61987]: DEBUG nova.network.neutron [req-f9cf16e0-1e64-4127-b7fd-0d727c0a4e0a req-1b4147f8-6a54-4b00-8c44-f69a86adf161 service nova] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Refreshing network info cache for port 5f60a42f-4af4-440f-93a1-79dacf25140a {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 596.248415] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-96615940-ff36-4d3e-b61b-1a377a84569d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.263548] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec997f3c-ebea-4201-b791-09fa7ba5bc9c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.292025] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 660e5e4c-0804-429b-ab18-9305a433391f could not be found. [ 596.292396] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 596.292706] env[61987]: INFO nova.compute.manager [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Took 0.05 seconds to destroy the instance on the hypervisor. [ 596.293088] env[61987]: DEBUG oslo.service.loopingcall [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 596.295771] env[61987]: DEBUG nova.compute.manager [-] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 596.295994] env[61987]: DEBUG nova.network.neutron [-] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 596.325087] env[61987]: DEBUG nova.network.neutron [-] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 596.562108] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98d65307-aa5c-4455-91bd-e5a543b20990 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.571085] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e537600f-e584-46a4-871a-65840a134491 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.610440] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fed84c6a-7052-4918-936d-de3c3142298c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.617454] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-569bc377-e0e3-4471-92a4-c18cf5bdad3d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.633293] env[61987]: DEBUG nova.compute.provider_tree [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.732509] env[61987]: DEBUG oslo_vmware.api [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061242, 'name': PowerOnVM_Task, 'duration_secs': 0.414652} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 596.732780] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 596.733017] env[61987]: INFO nova.compute.manager [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Took 10.06 seconds to spawn the instance on the hypervisor. [ 596.733287] env[61987]: DEBUG nova.compute.manager [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 596.734049] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed78f69-b2b2-4253-b2e4-eec65a3e4d75 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.777950] env[61987]: DEBUG nova.network.neutron [req-f9cf16e0-1e64-4127-b7fd-0d727c0a4e0a req-1b4147f8-6a54-4b00-8c44-f69a86adf161 service nova] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 596.827585] env[61987]: DEBUG nova.network.neutron [-] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.920324] env[61987]: DEBUG nova.network.neutron [req-f9cf16e0-1e64-4127-b7fd-0d727c0a4e0a req-1b4147f8-6a54-4b00-8c44-f69a86adf161 service nova] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.073430] env[61987]: ERROR nova.compute.manager [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 45aa1dda-6ba4-4927-9269-11fba215aa5a, please check neutron logs for more information. [ 597.073430] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 597.073430] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 597.073430] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 597.073430] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.073430] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 597.073430] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.073430] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 597.073430] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.073430] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 597.073430] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.073430] env[61987]: ERROR nova.compute.manager raise self.value [ 597.073430] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.073430] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 597.073430] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.073430] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 597.074134] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.074134] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 597.074134] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 45aa1dda-6ba4-4927-9269-11fba215aa5a, please check neutron logs for more information. [ 597.074134] env[61987]: ERROR nova.compute.manager [ 597.074134] env[61987]: Traceback (most recent call last): [ 597.074134] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 597.074134] env[61987]: listener.cb(fileno) [ 597.074134] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.074134] env[61987]: result = function(*args, **kwargs) [ 597.074134] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.074134] env[61987]: return func(*args, **kwargs) [ 597.074134] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 597.074134] env[61987]: raise e [ 597.074134] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 597.074134] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 597.074134] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.074134] env[61987]: created_port_ids = self._update_ports_for_instance( [ 597.074134] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.074134] env[61987]: with excutils.save_and_reraise_exception(): [ 597.074134] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.074134] env[61987]: self.force_reraise() [ 597.074134] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.074134] env[61987]: raise self.value [ 597.074134] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.074134] env[61987]: updated_port = self._update_port( [ 597.074134] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.074134] env[61987]: _ensure_no_port_binding_failure(port) [ 597.074134] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.074134] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 597.074945] env[61987]: nova.exception.PortBindingFailed: Binding failed for port 45aa1dda-6ba4-4927-9269-11fba215aa5a, please check neutron logs for more information. [ 597.074945] env[61987]: Removing descriptor: 19 [ 597.109995] env[61987]: DEBUG nova.compute.manager [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 597.135630] env[61987]: DEBUG nova.virt.hardware [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 597.135901] env[61987]: DEBUG nova.virt.hardware [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 597.137029] env[61987]: DEBUG nova.virt.hardware [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 597.137029] env[61987]: DEBUG nova.virt.hardware [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 597.137029] env[61987]: DEBUG nova.virt.hardware [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 597.137029] env[61987]: DEBUG nova.virt.hardware [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 597.137029] env[61987]: DEBUG nova.virt.hardware [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 597.137371] env[61987]: DEBUG nova.virt.hardware [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 597.137371] env[61987]: DEBUG nova.virt.hardware [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 597.137371] env[61987]: DEBUG nova.virt.hardware [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 597.137546] env[61987]: DEBUG nova.virt.hardware [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 597.138655] env[61987]: DEBUG nova.scheduler.client.report [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 597.142830] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31b47740-7544-4edf-947f-259673fec4bd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.150472] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd1ce72-7a7b-476f-8747-baf4059b0643 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.157348] env[61987]: DEBUG nova.compute.manager [req-88fb9f97-ee34-4fac-9ec9-9d9205eb3e16 req-720bd204-fed6-482a-b0ed-f213308e3ae3 service nova] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Received event network-changed-45aa1dda-6ba4-4927-9269-11fba215aa5a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 597.157470] env[61987]: DEBUG nova.compute.manager [req-88fb9f97-ee34-4fac-9ec9-9d9205eb3e16 req-720bd204-fed6-482a-b0ed-f213308e3ae3 service nova] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Refreshing instance network info cache due to event network-changed-45aa1dda-6ba4-4927-9269-11fba215aa5a. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 597.157680] env[61987]: DEBUG oslo_concurrency.lockutils [req-88fb9f97-ee34-4fac-9ec9-9d9205eb3e16 req-720bd204-fed6-482a-b0ed-f213308e3ae3 service nova] Acquiring lock "refresh_cache-c7bfab3b-bfb1-416f-a302-9956c4fb0748" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.157821] env[61987]: DEBUG oslo_concurrency.lockutils [req-88fb9f97-ee34-4fac-9ec9-9d9205eb3e16 req-720bd204-fed6-482a-b0ed-f213308e3ae3 service nova] Acquired lock "refresh_cache-c7bfab3b-bfb1-416f-a302-9956c4fb0748" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.157983] env[61987]: DEBUG nova.network.neutron [req-88fb9f97-ee34-4fac-9ec9-9d9205eb3e16 req-720bd204-fed6-482a-b0ed-f213308e3ae3 service nova] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Refreshing network info cache for port 45aa1dda-6ba4-4927-9269-11fba215aa5a {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 597.174485] env[61987]: ERROR nova.compute.manager [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 45aa1dda-6ba4-4927-9269-11fba215aa5a, please check neutron logs for more information. [ 597.174485] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Traceback (most recent call last): [ 597.174485] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 597.174485] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] yield resources [ 597.174485] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 597.174485] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] self.driver.spawn(context, instance, image_meta, [ 597.174485] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 597.174485] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.174485] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.174485] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] vm_ref = self.build_virtual_machine(instance, [ 597.174485] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.174884] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.174884] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.174884] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] for vif in network_info: [ 597.174884] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.174884] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] return self._sync_wrapper(fn, *args, **kwargs) [ 597.174884] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.174884] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] self.wait() [ 597.174884] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.174884] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] self[:] = self._gt.wait() [ 597.174884] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.174884] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] return self._exit_event.wait() [ 597.174884] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 597.174884] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] current.throw(*self._exc) [ 597.175281] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.175281] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] result = function(*args, **kwargs) [ 597.175281] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.175281] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] return func(*args, **kwargs) [ 597.175281] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 597.175281] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] raise e [ 597.175281] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 597.175281] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] nwinfo = self.network_api.allocate_for_instance( [ 597.175281] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.175281] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] created_port_ids = self._update_ports_for_instance( [ 597.175281] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.175281] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] with excutils.save_and_reraise_exception(): [ 597.175281] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.175677] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] self.force_reraise() [ 597.175677] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.175677] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] raise self.value [ 597.175677] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.175677] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] updated_port = self._update_port( [ 597.175677] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.175677] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] _ensure_no_port_binding_failure(port) [ 597.175677] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.175677] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] raise exception.PortBindingFailed(port_id=port['id']) [ 597.175677] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] nova.exception.PortBindingFailed: Binding failed for port 45aa1dda-6ba4-4927-9269-11fba215aa5a, please check neutron logs for more information. [ 597.175677] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] [ 597.175677] env[61987]: INFO nova.compute.manager [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Terminating instance [ 597.250787] env[61987]: INFO nova.compute.manager [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Took 40.76 seconds to build instance. [ 597.329585] env[61987]: INFO nova.compute.manager [-] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Took 1.03 seconds to deallocate network for instance. [ 597.332021] env[61987]: DEBUG nova.compute.claims [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 597.333013] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.423729] env[61987]: DEBUG oslo_concurrency.lockutils [req-f9cf16e0-1e64-4127-b7fd-0d727c0a4e0a req-1b4147f8-6a54-4b00-8c44-f69a86adf161 service nova] Releasing lock "refresh_cache-660e5e4c-0804-429b-ab18-9305a433391f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.424744] env[61987]: DEBUG nova.compute.manager [req-f9cf16e0-1e64-4127-b7fd-0d727c0a4e0a req-1b4147f8-6a54-4b00-8c44-f69a86adf161 service nova] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Received event network-vif-deleted-5f60a42f-4af4-440f-93a1-79dacf25140a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 597.562764] env[61987]: DEBUG nova.compute.manager [None req-f93932ba-dc5d-4331-bbde-338d971406d9 tempest-ServerDiagnosticsV248Test-203423251 tempest-ServerDiagnosticsV248Test-203423251-project-admin] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 597.563649] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c7f5f36-78ea-4d30-a0f7-ab85b56a0fe2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.570756] env[61987]: INFO nova.compute.manager [None req-f93932ba-dc5d-4331-bbde-338d971406d9 tempest-ServerDiagnosticsV248Test-203423251 tempest-ServerDiagnosticsV248Test-203423251-project-admin] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Retrieving diagnostics [ 597.571703] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22a8f750-c239-4c8e-b0a9-0ae1bc53cef1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.646170] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.559s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.646697] env[61987]: DEBUG nova.compute.manager [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 597.649116] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.884s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.676909] env[61987]: DEBUG nova.network.neutron [req-88fb9f97-ee34-4fac-9ec9-9d9205eb3e16 req-720bd204-fed6-482a-b0ed-f213308e3ae3 service nova] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.678699] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Acquiring lock "refresh_cache-c7bfab3b-bfb1-416f-a302-9956c4fb0748" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.752012] env[61987]: DEBUG oslo_concurrency.lockutils [None req-151f26e0-add1-49a9-a6d2-c030632c37ab tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Lock "6bf0d1f1-d079-4956-bc9a-86ba66885f3a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.331s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.905624] env[61987]: DEBUG nova.network.neutron [req-88fb9f97-ee34-4fac-9ec9-9d9205eb3e16 req-720bd204-fed6-482a-b0ed-f213308e3ae3 service nova] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.157128] env[61987]: DEBUG nova.compute.utils [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 598.160353] env[61987]: DEBUG nova.compute.manager [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 598.160353] env[61987]: DEBUG nova.network.neutron [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 598.212517] env[61987]: DEBUG nova.policy [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c124c6b758543a68e141796b585e41f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bd47eb44bc334bf3ae5813905903ecbc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 598.255187] env[61987]: DEBUG nova.compute.manager [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 598.409342] env[61987]: DEBUG oslo_concurrency.lockutils [req-88fb9f97-ee34-4fac-9ec9-9d9205eb3e16 req-720bd204-fed6-482a-b0ed-f213308e3ae3 service nova] Releasing lock "refresh_cache-c7bfab3b-bfb1-416f-a302-9956c4fb0748" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.409342] env[61987]: DEBUG nova.compute.manager [req-88fb9f97-ee34-4fac-9ec9-9d9205eb3e16 req-720bd204-fed6-482a-b0ed-f213308e3ae3 service nova] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Received event network-vif-deleted-45aa1dda-6ba4-4927-9269-11fba215aa5a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 598.409342] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Acquired lock "refresh_cache-c7bfab3b-bfb1-416f-a302-9956c4fb0748" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.409342] env[61987]: DEBUG nova.network.neutron [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 598.544593] env[61987]: DEBUG nova.network.neutron [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Successfully created port: 36e53101-e4f1-4e24-b454-a7cda6c93d8d {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 598.549565] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf769100-984f-4304-8139-521b8b2d3646 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.557205] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd60dda-4894-4c16-b1db-4af8b75ea00a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.586994] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f27e2265-d080-42d7-9cd7-202e61f4c714 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.594938] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d3055af-fafe-4f43-84f5-8e79ba2ceff6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.607963] env[61987]: DEBUG nova.compute.provider_tree [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.660521] env[61987]: DEBUG nova.compute.manager [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 598.774926] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.930984] env[61987]: DEBUG nova.network.neutron [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.050515] env[61987]: DEBUG nova.network.neutron [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.113019] env[61987]: DEBUG nova.scheduler.client.report [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 599.418705] env[61987]: DEBUG nova.compute.manager [req-2367deab-82e2-4816-9228-f804bcd84e5b req-05991632-52c8-482e-89fa-4b727744b485 service nova] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Received event network-changed-36e53101-e4f1-4e24-b454-a7cda6c93d8d {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 599.418981] env[61987]: DEBUG nova.compute.manager [req-2367deab-82e2-4816-9228-f804bcd84e5b req-05991632-52c8-482e-89fa-4b727744b485 service nova] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Refreshing instance network info cache due to event network-changed-36e53101-e4f1-4e24-b454-a7cda6c93d8d. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 599.419100] env[61987]: DEBUG oslo_concurrency.lockutils [req-2367deab-82e2-4816-9228-f804bcd84e5b req-05991632-52c8-482e-89fa-4b727744b485 service nova] Acquiring lock "refresh_cache-1915d6b8-2798-4e35-9a35-4191659c2cb6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.419246] env[61987]: DEBUG oslo_concurrency.lockutils [req-2367deab-82e2-4816-9228-f804bcd84e5b req-05991632-52c8-482e-89fa-4b727744b485 service nova] Acquired lock "refresh_cache-1915d6b8-2798-4e35-9a35-4191659c2cb6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.419403] env[61987]: DEBUG nova.network.neutron [req-2367deab-82e2-4816-9228-f804bcd84e5b req-05991632-52c8-482e-89fa-4b727744b485 service nova] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Refreshing network info cache for port 36e53101-e4f1-4e24-b454-a7cda6c93d8d {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 599.553442] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Releasing lock "refresh_cache-c7bfab3b-bfb1-416f-a302-9956c4fb0748" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.553986] env[61987]: DEBUG nova.compute.manager [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 599.554200] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 599.555164] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a1933c8d-15e9-4b92-934a-90ece39fefb4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.564660] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-265b6a6d-582c-4408-bf96-59aa6838b77b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.586516] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c7bfab3b-bfb1-416f-a302-9956c4fb0748 could not be found. [ 599.586731] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 599.586913] env[61987]: INFO nova.compute.manager [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Took 0.03 seconds to destroy the instance on the hypervisor. [ 599.587168] env[61987]: DEBUG oslo.service.loopingcall [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 599.587391] env[61987]: DEBUG nova.compute.manager [-] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 599.587483] env[61987]: DEBUG nova.network.neutron [-] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 599.606926] env[61987]: DEBUG nova.network.neutron [-] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.615917] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.967s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.616534] env[61987]: ERROR nova.compute.manager [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 94295cbb-b5de-4392-853a-6e75332a2ac4, please check neutron logs for more information. [ 599.616534] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Traceback (most recent call last): [ 599.616534] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 599.616534] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] self.driver.spawn(context, instance, image_meta, [ 599.616534] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 599.616534] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] self._vmops.spawn(context, instance, image_meta, injected_files, [ 599.616534] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 599.616534] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] vm_ref = self.build_virtual_machine(instance, [ 599.616534] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 599.616534] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] vif_infos = vmwarevif.get_vif_info(self._session, [ 599.616534] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 599.616895] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] for vif in network_info: [ 599.616895] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 599.616895] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] return self._sync_wrapper(fn, *args, **kwargs) [ 599.616895] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 599.616895] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] self.wait() [ 599.616895] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 599.616895] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] self[:] = self._gt.wait() [ 599.616895] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 599.616895] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] return self._exit_event.wait() [ 599.616895] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 599.616895] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] result = hub.switch() [ 599.616895] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 599.616895] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] return self.greenlet.switch() [ 599.617501] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.617501] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] result = function(*args, **kwargs) [ 599.617501] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 599.617501] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] return func(*args, **kwargs) [ 599.617501] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 599.617501] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] raise e [ 599.617501] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 599.617501] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] nwinfo = self.network_api.allocate_for_instance( [ 599.617501] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 599.617501] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] created_port_ids = self._update_ports_for_instance( [ 599.617501] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 599.617501] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] with excutils.save_and_reraise_exception(): [ 599.617501] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.618061] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] self.force_reraise() [ 599.618061] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.618061] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] raise self.value [ 599.618061] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 599.618061] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] updated_port = self._update_port( [ 599.618061] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.618061] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] _ensure_no_port_binding_failure(port) [ 599.618061] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.618061] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] raise exception.PortBindingFailed(port_id=port['id']) [ 599.618061] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] nova.exception.PortBindingFailed: Binding failed for port 94295cbb-b5de-4392-853a-6e75332a2ac4, please check neutron logs for more information. [ 599.618061] env[61987]: ERROR nova.compute.manager [instance: 3182ca24-c989-4c29-8e47-d913d0607240] [ 599.618555] env[61987]: DEBUG nova.compute.utils [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Binding failed for port 94295cbb-b5de-4392-853a-6e75332a2ac4, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 599.618555] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.600s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.621572] env[61987]: DEBUG nova.compute.manager [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Build of instance 3182ca24-c989-4c29-8e47-d913d0607240 was re-scheduled: Binding failed for port 94295cbb-b5de-4392-853a-6e75332a2ac4, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 599.622005] env[61987]: DEBUG nova.compute.manager [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 599.622240] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquiring lock "refresh_cache-3182ca24-c989-4c29-8e47-d913d0607240" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.622385] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquired lock "refresh_cache-3182ca24-c989-4c29-8e47-d913d0607240" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.622539] env[61987]: DEBUG nova.network.neutron [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 599.658175] env[61987]: ERROR nova.compute.manager [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 36e53101-e4f1-4e24-b454-a7cda6c93d8d, please check neutron logs for more information. [ 599.658175] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 599.658175] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 599.658175] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 599.658175] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 599.658175] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 599.658175] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 599.658175] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 599.658175] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.658175] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 599.658175] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.658175] env[61987]: ERROR nova.compute.manager raise self.value [ 599.658175] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 599.658175] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 599.658175] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.658175] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 599.661280] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.661280] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 599.661280] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 36e53101-e4f1-4e24-b454-a7cda6c93d8d, please check neutron logs for more information. [ 599.661280] env[61987]: ERROR nova.compute.manager [ 599.661280] env[61987]: Traceback (most recent call last): [ 599.661280] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 599.661280] env[61987]: listener.cb(fileno) [ 599.661280] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.661280] env[61987]: result = function(*args, **kwargs) [ 599.661280] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 599.661280] env[61987]: return func(*args, **kwargs) [ 599.661280] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 599.661280] env[61987]: raise e [ 599.661280] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 599.661280] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 599.661280] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 599.661280] env[61987]: created_port_ids = self._update_ports_for_instance( [ 599.661280] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 599.661280] env[61987]: with excutils.save_and_reraise_exception(): [ 599.661280] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.661280] env[61987]: self.force_reraise() [ 599.661280] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.661280] env[61987]: raise self.value [ 599.661280] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 599.661280] env[61987]: updated_port = self._update_port( [ 599.661280] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.661280] env[61987]: _ensure_no_port_binding_failure(port) [ 599.661280] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.661280] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 599.662340] env[61987]: nova.exception.PortBindingFailed: Binding failed for port 36e53101-e4f1-4e24-b454-a7cda6c93d8d, please check neutron logs for more information. [ 599.662340] env[61987]: Removing descriptor: 19 [ 599.674823] env[61987]: DEBUG nova.compute.manager [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 599.738090] env[61987]: DEBUG nova.virt.hardware [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 599.738354] env[61987]: DEBUG nova.virt.hardware [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 599.738513] env[61987]: DEBUG nova.virt.hardware [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 599.738696] env[61987]: DEBUG nova.virt.hardware [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 599.738845] env[61987]: DEBUG nova.virt.hardware [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 599.738994] env[61987]: DEBUG nova.virt.hardware [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 599.739280] env[61987]: DEBUG nova.virt.hardware [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 599.739440] env[61987]: DEBUG nova.virt.hardware [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 599.739611] env[61987]: DEBUG nova.virt.hardware [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 599.739785] env[61987]: DEBUG nova.virt.hardware [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 599.739976] env[61987]: DEBUG nova.virt.hardware [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 599.740870] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6c656fe-0028-4589-84f2-575c8b3f5a6a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.749963] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5e2a97e-9592-404f-ac38-748adb8c4d2d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.765396] env[61987]: ERROR nova.compute.manager [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 36e53101-e4f1-4e24-b454-a7cda6c93d8d, please check neutron logs for more information. [ 599.765396] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Traceback (most recent call last): [ 599.765396] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 599.765396] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] yield resources [ 599.765396] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 599.765396] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] self.driver.spawn(context, instance, image_meta, [ 599.765396] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 599.765396] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 599.765396] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 599.765396] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] vm_ref = self.build_virtual_machine(instance, [ 599.765396] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 599.765696] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] vif_infos = vmwarevif.get_vif_info(self._session, [ 599.765696] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 599.765696] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] for vif in network_info: [ 599.765696] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 599.765696] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] return self._sync_wrapper(fn, *args, **kwargs) [ 599.765696] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 599.765696] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] self.wait() [ 599.765696] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 599.765696] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] self[:] = self._gt.wait() [ 599.765696] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 599.765696] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] return self._exit_event.wait() [ 599.765696] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 599.765696] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] current.throw(*self._exc) [ 599.766041] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 599.766041] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] result = function(*args, **kwargs) [ 599.766041] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 599.766041] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] return func(*args, **kwargs) [ 599.766041] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 599.766041] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] raise e [ 599.766041] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 599.766041] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] nwinfo = self.network_api.allocate_for_instance( [ 599.766041] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 599.766041] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] created_port_ids = self._update_ports_for_instance( [ 599.766041] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 599.766041] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] with excutils.save_and_reraise_exception(): [ 599.766041] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 599.766344] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] self.force_reraise() [ 599.766344] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 599.766344] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] raise self.value [ 599.766344] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 599.766344] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] updated_port = self._update_port( [ 599.766344] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 599.766344] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] _ensure_no_port_binding_failure(port) [ 599.766344] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 599.766344] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] raise exception.PortBindingFailed(port_id=port['id']) [ 599.766344] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] nova.exception.PortBindingFailed: Binding failed for port 36e53101-e4f1-4e24-b454-a7cda6c93d8d, please check neutron logs for more information. [ 599.766344] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] [ 599.766344] env[61987]: INFO nova.compute.manager [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Terminating instance [ 599.940807] env[61987]: DEBUG nova.network.neutron [req-2367deab-82e2-4816-9228-f804bcd84e5b req-05991632-52c8-482e-89fa-4b727744b485 service nova] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 600.024352] env[61987]: DEBUG nova.network.neutron [req-2367deab-82e2-4816-9228-f804bcd84e5b req-05991632-52c8-482e-89fa-4b727744b485 service nova] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.109658] env[61987]: DEBUG nova.network.neutron [-] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.146595] env[61987]: DEBUG nova.network.neutron [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 600.194558] env[61987]: DEBUG nova.network.neutron [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.270296] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "refresh_cache-1915d6b8-2798-4e35-9a35-4191659c2cb6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.495331] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48e5ed2-5b6e-4491-bc5b-9c628548e9d0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.502897] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa0baafe-f0f7-4ed0-ad83-beef209e6b45 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.532522] env[61987]: DEBUG oslo_concurrency.lockutils [req-2367deab-82e2-4816-9228-f804bcd84e5b req-05991632-52c8-482e-89fa-4b727744b485 service nova] Releasing lock "refresh_cache-1915d6b8-2798-4e35-9a35-4191659c2cb6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.533201] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired lock "refresh_cache-1915d6b8-2798-4e35-9a35-4191659c2cb6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.533331] env[61987]: DEBUG nova.network.neutron [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 600.534826] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4624f71f-9a1a-467b-b52d-7f7557dc36c0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.542719] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66c7d00d-79a1-49ce-abe9-d6250a140c80 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.556192] env[61987]: DEBUG nova.compute.provider_tree [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 600.612045] env[61987]: INFO nova.compute.manager [-] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Took 1.02 seconds to deallocate network for instance. [ 600.614216] env[61987]: DEBUG nova.compute.claims [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 600.615057] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.697054] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Releasing lock "refresh_cache-3182ca24-c989-4c29-8e47-d913d0607240" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.697393] env[61987]: DEBUG nova.compute.manager [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 600.697615] env[61987]: DEBUG nova.compute.manager [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 600.697855] env[61987]: DEBUG nova.network.neutron [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 600.711392] env[61987]: DEBUG nova.network.neutron [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.054332] env[61987]: DEBUG nova.network.neutron [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.059020] env[61987]: DEBUG nova.scheduler.client.report [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 601.136478] env[61987]: DEBUG nova.network.neutron [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.214134] env[61987]: DEBUG nova.network.neutron [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.453573] env[61987]: DEBUG nova.compute.manager [req-fa44139a-c1a1-4022-b154-f0277fb039df req-d52e241b-c012-4137-8689-c1d5c02ef378 service nova] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Received event network-vif-deleted-36e53101-e4f1-4e24-b454-a7cda6c93d8d {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 601.563428] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.945s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.564136] env[61987]: ERROR nova.compute.manager [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 36e6b6a6-fca0-41c9-8284-b06c55e23a01, please check neutron logs for more information. [ 601.564136] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Traceback (most recent call last): [ 601.564136] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 601.564136] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] self.driver.spawn(context, instance, image_meta, [ 601.564136] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 601.564136] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.564136] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.564136] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] vm_ref = self.build_virtual_machine(instance, [ 601.564136] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.564136] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.564136] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.564394] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] for vif in network_info: [ 601.564394] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.564394] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] return self._sync_wrapper(fn, *args, **kwargs) [ 601.564394] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.564394] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] self.wait() [ 601.564394] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.564394] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] self[:] = self._gt.wait() [ 601.564394] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.564394] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] return self._exit_event.wait() [ 601.564394] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.564394] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] result = hub.switch() [ 601.564394] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.564394] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] return self.greenlet.switch() [ 601.564674] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.564674] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] result = function(*args, **kwargs) [ 601.564674] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.564674] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] return func(*args, **kwargs) [ 601.564674] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 601.564674] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] raise e [ 601.564674] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 601.564674] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] nwinfo = self.network_api.allocate_for_instance( [ 601.564674] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 601.564674] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] created_port_ids = self._update_ports_for_instance( [ 601.564674] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 601.564674] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] with excutils.save_and_reraise_exception(): [ 601.564674] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.565265] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] self.force_reraise() [ 601.565265] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.565265] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] raise self.value [ 601.565265] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 601.565265] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] updated_port = self._update_port( [ 601.565265] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.565265] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] _ensure_no_port_binding_failure(port) [ 601.565265] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.565265] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] raise exception.PortBindingFailed(port_id=port['id']) [ 601.565265] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] nova.exception.PortBindingFailed: Binding failed for port 36e6b6a6-fca0-41c9-8284-b06c55e23a01, please check neutron logs for more information. [ 601.565265] env[61987]: ERROR nova.compute.manager [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] [ 601.565676] env[61987]: DEBUG nova.compute.utils [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Binding failed for port 36e6b6a6-fca0-41c9-8284-b06c55e23a01, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 601.566057] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.011s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.568744] env[61987]: DEBUG nova.compute.manager [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Build of instance 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7 was re-scheduled: Binding failed for port 36e6b6a6-fca0-41c9-8284-b06c55e23a01, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 601.569229] env[61987]: DEBUG nova.compute.manager [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 601.569457] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Acquiring lock "refresh_cache-7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.569741] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Acquired lock "refresh_cache-7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.569857] env[61987]: DEBUG nova.network.neutron [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 601.639305] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Releasing lock "refresh_cache-1915d6b8-2798-4e35-9a35-4191659c2cb6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.639730] env[61987]: DEBUG nova.compute.manager [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 601.639922] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 601.640221] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5ff6c566-8d0e-41a0-baeb-073bf73463ee {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.650197] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38aef206-9df2-4923-97a9-771ece979d67 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.671524] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1915d6b8-2798-4e35-9a35-4191659c2cb6 could not be found. [ 601.671785] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 601.671980] env[61987]: INFO nova.compute.manager [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Took 0.03 seconds to destroy the instance on the hypervisor. [ 601.672260] env[61987]: DEBUG oslo.service.loopingcall [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 601.672484] env[61987]: DEBUG nova.compute.manager [-] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 601.672579] env[61987]: DEBUG nova.network.neutron [-] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 601.693168] env[61987]: DEBUG nova.network.neutron [-] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.717197] env[61987]: INFO nova.compute.manager [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 3182ca24-c989-4c29-8e47-d913d0607240] Took 1.02 seconds to deallocate network for instance. [ 602.089875] env[61987]: DEBUG nova.network.neutron [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 602.177415] env[61987]: DEBUG nova.network.neutron [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.195694] env[61987]: DEBUG nova.network.neutron [-] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.424844] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12080863-9df8-45f9-8f23-43e183b9e14b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.432793] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109a1d7f-0671-46aa-9f05-294a184ddeaa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.462571] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-947c9fe6-193d-4cf5-ac74-3b6ca265f00c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.469961] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-827cd55b-8b17-452e-adb8-64e4976339d1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.485058] env[61987]: DEBUG nova.compute.provider_tree [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 602.680449] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Releasing lock "refresh_cache-7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.680766] env[61987]: DEBUG nova.compute.manager [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 602.680877] env[61987]: DEBUG nova.compute.manager [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 602.681102] env[61987]: DEBUG nova.network.neutron [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 602.694553] env[61987]: DEBUG nova.network.neutron [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 602.697424] env[61987]: INFO nova.compute.manager [-] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Took 1.02 seconds to deallocate network for instance. [ 602.699721] env[61987]: DEBUG nova.compute.claims [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 602.699891] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.742458] env[61987]: INFO nova.scheduler.client.report [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Deleted allocations for instance 3182ca24-c989-4c29-8e47-d913d0607240 [ 602.988642] env[61987]: DEBUG nova.scheduler.client.report [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 603.197013] env[61987]: DEBUG nova.network.neutron [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.250512] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5f273f35-4f1c-4eb0-882c-35756bd96541 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "3182ca24-c989-4c29-8e47-d913d0607240" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.448s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.493883] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.928s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.494614] env[61987]: ERROR nova.compute.manager [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 53be8281-4dc3-4ab6-abb1-a65876f596ef, please check neutron logs for more information. [ 603.494614] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Traceback (most recent call last): [ 603.494614] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 603.494614] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] self.driver.spawn(context, instance, image_meta, [ 603.494614] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 603.494614] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 603.494614] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 603.494614] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] vm_ref = self.build_virtual_machine(instance, [ 603.494614] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 603.494614] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] vif_infos = vmwarevif.get_vif_info(self._session, [ 603.494614] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 603.494869] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] for vif in network_info: [ 603.494869] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 603.494869] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] return self._sync_wrapper(fn, *args, **kwargs) [ 603.494869] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 603.494869] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] self.wait() [ 603.494869] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 603.494869] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] self[:] = self._gt.wait() [ 603.494869] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 603.494869] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] return self._exit_event.wait() [ 603.494869] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 603.494869] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] result = hub.switch() [ 603.494869] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 603.494869] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] return self.greenlet.switch() [ 603.495144] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.495144] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] result = function(*args, **kwargs) [ 603.495144] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 603.495144] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] return func(*args, **kwargs) [ 603.495144] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 603.495144] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] raise e [ 603.495144] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 603.495144] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] nwinfo = self.network_api.allocate_for_instance( [ 603.495144] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 603.495144] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] created_port_ids = self._update_ports_for_instance( [ 603.495144] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 603.495144] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] with excutils.save_and_reraise_exception(): [ 603.495144] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.495438] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] self.force_reraise() [ 603.495438] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.495438] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] raise self.value [ 603.495438] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 603.495438] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] updated_port = self._update_port( [ 603.495438] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.495438] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] _ensure_no_port_binding_failure(port) [ 603.495438] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.495438] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] raise exception.PortBindingFailed(port_id=port['id']) [ 603.495438] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] nova.exception.PortBindingFailed: Binding failed for port 53be8281-4dc3-4ab6-abb1-a65876f596ef, please check neutron logs for more information. [ 603.495438] env[61987]: ERROR nova.compute.manager [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] [ 603.495741] env[61987]: DEBUG nova.compute.utils [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Binding failed for port 53be8281-4dc3-4ab6-abb1-a65876f596ef, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 603.497075] env[61987]: DEBUG nova.compute.manager [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Build of instance 7a94f776-579d-41e8-94aa-6095b3fd69c6 was re-scheduled: Binding failed for port 53be8281-4dc3-4ab6-abb1-a65876f596ef, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 603.497499] env[61987]: DEBUG nova.compute.manager [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 603.497730] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Acquiring lock "refresh_cache-7a94f776-579d-41e8-94aa-6095b3fd69c6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.497873] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Acquired lock "refresh_cache-7a94f776-579d-41e8-94aa-6095b3fd69c6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.498045] env[61987]: DEBUG nova.network.neutron [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 603.499700] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.013s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.702095] env[61987]: INFO nova.compute.manager [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7] Took 1.02 seconds to deallocate network for instance. [ 603.753474] env[61987]: DEBUG nova.compute.manager [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 604.018850] env[61987]: DEBUG nova.network.neutron [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.103822] env[61987]: DEBUG nova.network.neutron [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.270945] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.372926] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c001c0b-d6f8-4592-8fec-6a8b379d2480 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.381591] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aa52424-2345-468a-a93a-52157bf3cdea {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.413429] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e543e55-2773-4a60-8866-55c9557720ae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.420772] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb656372-09e3-42a9-b603-1cfa4280fe8e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.428528] env[61987]: DEBUG oslo_concurrency.lockutils [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquiring lock "80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.428761] env[61987]: DEBUG oslo_concurrency.lockutils [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.437623] env[61987]: DEBUG nova.compute.provider_tree [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 604.608469] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Releasing lock "refresh_cache-7a94f776-579d-41e8-94aa-6095b3fd69c6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.608614] env[61987]: DEBUG nova.compute.manager [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 604.608801] env[61987]: DEBUG nova.compute.manager [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 604.608970] env[61987]: DEBUG nova.network.neutron [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 604.622624] env[61987]: DEBUG nova.network.neutron [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.729619] env[61987]: INFO nova.scheduler.client.report [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Deleted allocations for instance 7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7 [ 604.940195] env[61987]: DEBUG nova.scheduler.client.report [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 605.125764] env[61987]: DEBUG nova.network.neutron [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.238603] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2f7c0870-db9c-49a2-971b-1b2f35a89e2f tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Lock "7a8c0f03-0bb9-4759-b8fb-97e82e7a69b7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.715s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.446014] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.946s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.446672] env[61987]: ERROR nova.compute.manager [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1aae5742-c738-447d-b60d-05f7c8f67871, please check neutron logs for more information. [ 605.446672] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Traceback (most recent call last): [ 605.446672] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 605.446672] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] self.driver.spawn(context, instance, image_meta, [ 605.446672] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 605.446672] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.446672] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.446672] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] vm_ref = self.build_virtual_machine(instance, [ 605.446672] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.446672] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.446672] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.447054] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] for vif in network_info: [ 605.447054] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 605.447054] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] return self._sync_wrapper(fn, *args, **kwargs) [ 605.447054] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 605.447054] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] self.wait() [ 605.447054] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 605.447054] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] self[:] = self._gt.wait() [ 605.447054] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.447054] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] return self._exit_event.wait() [ 605.447054] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 605.447054] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] result = hub.switch() [ 605.447054] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 605.447054] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] return self.greenlet.switch() [ 605.447423] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.447423] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] result = function(*args, **kwargs) [ 605.447423] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.447423] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] return func(*args, **kwargs) [ 605.447423] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 605.447423] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] raise e [ 605.447423] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 605.447423] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] nwinfo = self.network_api.allocate_for_instance( [ 605.447423] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 605.447423] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] created_port_ids = self._update_ports_for_instance( [ 605.447423] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 605.447423] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] with excutils.save_and_reraise_exception(): [ 605.447423] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.447801] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] self.force_reraise() [ 605.447801] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.447801] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] raise self.value [ 605.447801] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 605.447801] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] updated_port = self._update_port( [ 605.447801] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.447801] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] _ensure_no_port_binding_failure(port) [ 605.447801] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.447801] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] raise exception.PortBindingFailed(port_id=port['id']) [ 605.447801] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] nova.exception.PortBindingFailed: Binding failed for port 1aae5742-c738-447d-b60d-05f7c8f67871, please check neutron logs for more information. [ 605.447801] env[61987]: ERROR nova.compute.manager [instance: 938dab0c-9b30-404b-bca2-a1280a252584] [ 605.448157] env[61987]: DEBUG nova.compute.utils [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Binding failed for port 1aae5742-c738-447d-b60d-05f7c8f67871, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 605.449130] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.080s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.452761] env[61987]: DEBUG nova.compute.manager [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Build of instance 938dab0c-9b30-404b-bca2-a1280a252584 was re-scheduled: Binding failed for port 1aae5742-c738-447d-b60d-05f7c8f67871, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 605.453191] env[61987]: DEBUG nova.compute.manager [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 605.453418] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Acquiring lock "refresh_cache-938dab0c-9b30-404b-bca2-a1280a252584" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.453566] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Acquired lock "refresh_cache-938dab0c-9b30-404b-bca2-a1280a252584" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.453727] env[61987]: DEBUG nova.network.neutron [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 605.629340] env[61987]: INFO nova.compute.manager [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] [instance: 7a94f776-579d-41e8-94aa-6095b3fd69c6] Took 1.02 seconds to deallocate network for instance. [ 605.743025] env[61987]: DEBUG nova.compute.manager [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 605.986022] env[61987]: DEBUG nova.network.neutron [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.034075] env[61987]: DEBUG nova.network.neutron [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.265023] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.349345] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c23c6851-4ddf-429d-9495-9f141adc1f33 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.356985] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba76b08a-9c49-4a76-b180-714a87392741 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.389041] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63769273-148d-4a36-8d05-cdac4a795e15 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.396788] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f456dec-2d2f-4b2b-a30a-ad2ceb51cca6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.410145] env[61987]: DEBUG nova.compute.provider_tree [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.537216] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Releasing lock "refresh_cache-938dab0c-9b30-404b-bca2-a1280a252584" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.537487] env[61987]: DEBUG nova.compute.manager [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 606.537644] env[61987]: DEBUG nova.compute.manager [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 606.537812] env[61987]: DEBUG nova.network.neutron [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 606.555350] env[61987]: DEBUG nova.network.neutron [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 606.663129] env[61987]: INFO nova.scheduler.client.report [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Deleted allocations for instance 7a94f776-579d-41e8-94aa-6095b3fd69c6 [ 606.913322] env[61987]: DEBUG nova.scheduler.client.report [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 607.057597] env[61987]: DEBUG nova.network.neutron [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.170536] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f37312fd-602d-4a2b-9792-bf722450ad2d tempest-InstanceActionsTestJSON-743204214 tempest-InstanceActionsTestJSON-743204214-project-member] Lock "7a94f776-579d-41e8-94aa-6095b3fd69c6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.163s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.418956] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.970s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.419638] env[61987]: ERROR nova.compute.manager [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f516ce51-3f8d-4bf0-a708-95b2c95837f4, please check neutron logs for more information. [ 607.419638] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Traceback (most recent call last): [ 607.419638] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 607.419638] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] self.driver.spawn(context, instance, image_meta, [ 607.419638] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 607.419638] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] self._vmops.spawn(context, instance, image_meta, injected_files, [ 607.419638] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 607.419638] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] vm_ref = self.build_virtual_machine(instance, [ 607.419638] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 607.419638] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] vif_infos = vmwarevif.get_vif_info(self._session, [ 607.419638] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 607.419908] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] for vif in network_info: [ 607.419908] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 607.419908] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] return self._sync_wrapper(fn, *args, **kwargs) [ 607.419908] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 607.419908] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] self.wait() [ 607.419908] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 607.419908] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] self[:] = self._gt.wait() [ 607.419908] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 607.419908] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] return self._exit_event.wait() [ 607.419908] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 607.419908] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] current.throw(*self._exc) [ 607.419908] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.419908] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] result = function(*args, **kwargs) [ 607.420407] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.420407] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] return func(*args, **kwargs) [ 607.420407] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 607.420407] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] raise e [ 607.420407] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 607.420407] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] nwinfo = self.network_api.allocate_for_instance( [ 607.420407] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 607.420407] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] created_port_ids = self._update_ports_for_instance( [ 607.420407] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 607.420407] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] with excutils.save_and_reraise_exception(): [ 607.420407] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.420407] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] self.force_reraise() [ 607.420407] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.420743] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] raise self.value [ 607.420743] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 607.420743] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] updated_port = self._update_port( [ 607.420743] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.420743] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] _ensure_no_port_binding_failure(port) [ 607.420743] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.420743] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] raise exception.PortBindingFailed(port_id=port['id']) [ 607.420743] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] nova.exception.PortBindingFailed: Binding failed for port f516ce51-3f8d-4bf0-a708-95b2c95837f4, please check neutron logs for more information. [ 607.420743] env[61987]: ERROR nova.compute.manager [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] [ 607.420743] env[61987]: DEBUG nova.compute.utils [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Binding failed for port f516ce51-3f8d-4bf0-a708-95b2c95837f4, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 607.421657] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.445s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.425918] env[61987]: DEBUG nova.compute.manager [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Build of instance 7961bd0d-de24-4e13-94a3-79c48c593813 was re-scheduled: Binding failed for port f516ce51-3f8d-4bf0-a708-95b2c95837f4, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 607.426437] env[61987]: DEBUG nova.compute.manager [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 607.427734] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Acquiring lock "refresh_cache-7961bd0d-de24-4e13-94a3-79c48c593813" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.427734] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Acquired lock "refresh_cache-7961bd0d-de24-4e13-94a3-79c48c593813" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.427734] env[61987]: DEBUG nova.network.neutron [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 607.512539] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Acquiring lock "93174a10-f8b1-4789-ab3c-dda07bdddaa8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.512787] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Lock "93174a10-f8b1-4789-ab3c-dda07bdddaa8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.560514] env[61987]: INFO nova.compute.manager [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] [instance: 938dab0c-9b30-404b-bca2-a1280a252584] Took 1.02 seconds to deallocate network for instance. [ 607.673902] env[61987]: DEBUG nova.compute.manager [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 607.966743] env[61987]: DEBUG nova.network.neutron [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.064175] env[61987]: DEBUG nova.network.neutron [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.158143] env[61987]: DEBUG nova.compute.manager [None req-6ca547b3-593a-431d-a223-e407b8531e00 tempest-ServerDiagnosticsV248Test-203423251 tempest-ServerDiagnosticsV248Test-203423251-project-admin] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 608.161433] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d716a21-68b0-4812-8f3b-b5306a2d3669 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.168523] env[61987]: INFO nova.compute.manager [None req-6ca547b3-593a-431d-a223-e407b8531e00 tempest-ServerDiagnosticsV248Test-203423251 tempest-ServerDiagnosticsV248Test-203423251-project-admin] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Retrieving diagnostics [ 608.169473] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc1f8098-a7a9-4d96-bee3-a978e2852de2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.218179] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.394337] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbfc6dea-af8d-4a10-bf8a-5d75ebe68316 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.402101] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9542d891-bed1-40cc-a38c-5e3e13115c15 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.431510] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa19bff6-d894-46bd-94cf-4cf7c9c5e316 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.438609] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130e750b-458b-416f-aca3-31e099724f34 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.452319] env[61987]: DEBUG nova.compute.provider_tree [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.570405] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Releasing lock "refresh_cache-7961bd0d-de24-4e13-94a3-79c48c593813" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.570699] env[61987]: DEBUG nova.compute.manager [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 608.570896] env[61987]: DEBUG nova.compute.manager [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 608.571115] env[61987]: DEBUG nova.network.neutron [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 608.592995] env[61987]: DEBUG nova.network.neutron [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.601719] env[61987]: INFO nova.scheduler.client.report [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Deleted allocations for instance 938dab0c-9b30-404b-bca2-a1280a252584 [ 609.416305] env[61987]: DEBUG nova.scheduler.client.report [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 609.419723] env[61987]: DEBUG nova.network.neutron [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.420344] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1ac0d393-b462-4d2a-9d9c-502a34e5fb32 tempest-ServerExternalEventsTest-1175726155 tempest-ServerExternalEventsTest-1175726155-project-member] Lock "938dab0c-9b30-404b-bca2-a1280a252584" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.751s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.420638] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Acquiring lock "6bf0d1f1-d079-4956-bc9a-86ba66885f3a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.421035] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Lock "6bf0d1f1-d079-4956-bc9a-86ba66885f3a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.421258] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Acquiring lock "6bf0d1f1-d079-4956-bc9a-86ba66885f3a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.421446] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Lock "6bf0d1f1-d079-4956-bc9a-86ba66885f3a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.421604] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Lock "6bf0d1f1-d079-4956-bc9a-86ba66885f3a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.424438] env[61987]: INFO nova.compute.manager [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Terminating instance [ 609.924208] env[61987]: INFO nova.compute.manager [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] [instance: 7961bd0d-de24-4e13-94a3-79c48c593813] Took 1.35 seconds to deallocate network for instance. [ 609.928255] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.506s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.929968] env[61987]: ERROR nova.compute.manager [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e99cf30c-90d8-4653-9d2c-4abd600020bd, please check neutron logs for more information. [ 609.929968] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Traceback (most recent call last): [ 609.929968] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 609.929968] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] self.driver.spawn(context, instance, image_meta, [ 609.929968] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 609.929968] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.929968] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.929968] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] vm_ref = self.build_virtual_machine(instance, [ 609.929968] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.929968] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.929968] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.930293] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] for vif in network_info: [ 609.930293] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.930293] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] return self._sync_wrapper(fn, *args, **kwargs) [ 609.930293] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.930293] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] self.wait() [ 609.930293] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.930293] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] self[:] = self._gt.wait() [ 609.930293] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.930293] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] return self._exit_event.wait() [ 609.930293] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 609.930293] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] result = hub.switch() [ 609.930293] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 609.930293] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] return self.greenlet.switch() [ 609.930625] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.930625] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] result = function(*args, **kwargs) [ 609.930625] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.930625] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] return func(*args, **kwargs) [ 609.930625] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 609.930625] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] raise e [ 609.930625] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 609.930625] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] nwinfo = self.network_api.allocate_for_instance( [ 609.930625] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.930625] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] created_port_ids = self._update_ports_for_instance( [ 609.930625] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.930625] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] with excutils.save_and_reraise_exception(): [ 609.930625] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.930913] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] self.force_reraise() [ 609.930913] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.930913] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] raise self.value [ 609.930913] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.930913] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] updated_port = self._update_port( [ 609.930913] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.930913] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] _ensure_no_port_binding_failure(port) [ 609.930913] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.930913] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] raise exception.PortBindingFailed(port_id=port['id']) [ 609.930913] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] nova.exception.PortBindingFailed: Binding failed for port e99cf30c-90d8-4653-9d2c-4abd600020bd, please check neutron logs for more information. [ 609.930913] env[61987]: ERROR nova.compute.manager [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] [ 609.931175] env[61987]: DEBUG nova.compute.utils [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Binding failed for port e99cf30c-90d8-4653-9d2c-4abd600020bd, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 609.933885] env[61987]: DEBUG nova.compute.manager [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 609.938214] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.606s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.942576] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Acquiring lock "refresh_cache-6bf0d1f1-d079-4956-bc9a-86ba66885f3a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.942625] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Acquired lock "refresh_cache-6bf0d1f1-d079-4956-bc9a-86ba66885f3a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.943140] env[61987]: DEBUG nova.network.neutron [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 609.947162] env[61987]: DEBUG nova.compute.manager [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Build of instance bac5a3bc-5a58-4c1d-abb8-1aaf6151358a was re-scheduled: Binding failed for port e99cf30c-90d8-4653-9d2c-4abd600020bd, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 609.947162] env[61987]: DEBUG nova.compute.manager [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 609.947438] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Acquiring lock "refresh_cache-bac5a3bc-5a58-4c1d-abb8-1aaf6151358a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.947983] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Acquired lock "refresh_cache-bac5a3bc-5a58-4c1d-abb8-1aaf6151358a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.947983] env[61987]: DEBUG nova.network.neutron [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 610.474662] env[61987]: DEBUG oslo_concurrency.lockutils [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.494822] env[61987]: DEBUG nova.network.neutron [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.496921] env[61987]: DEBUG nova.network.neutron [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.596966] env[61987]: DEBUG nova.network.neutron [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.628703] env[61987]: DEBUG nova.network.neutron [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.876064] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-325fff89-fa11-4ffb-a842-7a8fbd77f983 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.883933] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf26758-4e12-41e5-b8e3-9252c812037d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.916050] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d90210b-48b9-4b13-86de-46bde5360c7a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.924155] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42753e3e-de90-4756-828b-d95e1d682549 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.946081] env[61987]: DEBUG nova.compute.provider_tree [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.977054] env[61987]: INFO nova.scheduler.client.report [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Deleted allocations for instance 7961bd0d-de24-4e13-94a3-79c48c593813 [ 611.103631] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Releasing lock "refresh_cache-6bf0d1f1-d079-4956-bc9a-86ba66885f3a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.104144] env[61987]: DEBUG nova.compute.manager [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 611.104378] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 611.105551] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ac7b1e-3d73-4431-9782-b86618bc1690 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.113642] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 611.113902] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-89580f95-b07d-4f90-a700-0aff38cbc877 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.123051] env[61987]: DEBUG oslo_vmware.api [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Waiting for the task: (returnval){ [ 611.123051] env[61987]: value = "task-1061243" [ 611.123051] env[61987]: _type = "Task" [ 611.123051] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.128261] env[61987]: DEBUG oslo_vmware.api [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061243, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.131846] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Releasing lock "refresh_cache-bac5a3bc-5a58-4c1d-abb8-1aaf6151358a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.132076] env[61987]: DEBUG nova.compute.manager [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 611.132258] env[61987]: DEBUG nova.compute.manager [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 611.132421] env[61987]: DEBUG nova.network.neutron [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 611.166792] env[61987]: DEBUG nova.network.neutron [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.451065] env[61987]: DEBUG nova.scheduler.client.report [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 611.484234] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0cc4503a-43e5-426e-8476-f7e165d3154d tempest-VolumesAssistedSnapshotsTest-1812334981 tempest-VolumesAssistedSnapshotsTest-1812334981-project-member] Lock "7961bd0d-de24-4e13-94a3-79c48c593813" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.300s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.632465] env[61987]: DEBUG oslo_vmware.api [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061243, 'name': PowerOffVM_Task, 'duration_secs': 0.184921} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 611.632732] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 611.632897] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 611.633180] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9cd1a4c4-0377-494e-8a2f-85f54347bd44 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.660025] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 611.661289] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 611.661548] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Deleting the datastore file [datastore2] 6bf0d1f1-d079-4956-bc9a-86ba66885f3a {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 611.661857] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-889c4968-6361-4f11-a925-692f8f002414 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.669160] env[61987]: DEBUG nova.network.neutron [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.673789] env[61987]: DEBUG oslo_vmware.api [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Waiting for the task: (returnval){ [ 611.673789] env[61987]: value = "task-1061245" [ 611.673789] env[61987]: _type = "Task" [ 611.673789] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 611.682153] env[61987]: DEBUG oslo_vmware.api [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061245, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 611.957331] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.019s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.960229] env[61987]: ERROR nova.compute.manager [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5f60a42f-4af4-440f-93a1-79dacf25140a, please check neutron logs for more information. [ 611.960229] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Traceback (most recent call last): [ 611.960229] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 611.960229] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] self.driver.spawn(context, instance, image_meta, [ 611.960229] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 611.960229] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.960229] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.960229] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] vm_ref = self.build_virtual_machine(instance, [ 611.960229] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.960229] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.960229] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.960497] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] for vif in network_info: [ 611.960497] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.960497] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] return self._sync_wrapper(fn, *args, **kwargs) [ 611.960497] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.960497] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] self.wait() [ 611.960497] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.960497] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] self[:] = self._gt.wait() [ 611.960497] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.960497] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] return self._exit_event.wait() [ 611.960497] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 611.960497] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] current.throw(*self._exc) [ 611.960497] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.960497] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] result = function(*args, **kwargs) [ 611.960817] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 611.960817] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] return func(*args, **kwargs) [ 611.960817] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 611.960817] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] raise e [ 611.960817] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 611.960817] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] nwinfo = self.network_api.allocate_for_instance( [ 611.960817] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 611.960817] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] created_port_ids = self._update_ports_for_instance( [ 611.960817] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 611.960817] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] with excutils.save_and_reraise_exception(): [ 611.960817] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.960817] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] self.force_reraise() [ 611.960817] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.961191] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] raise self.value [ 611.961191] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 611.961191] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] updated_port = self._update_port( [ 611.961191] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.961191] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] _ensure_no_port_binding_failure(port) [ 611.961191] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.961191] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] raise exception.PortBindingFailed(port_id=port['id']) [ 611.961191] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] nova.exception.PortBindingFailed: Binding failed for port 5f60a42f-4af4-440f-93a1-79dacf25140a, please check neutron logs for more information. [ 611.961191] env[61987]: ERROR nova.compute.manager [instance: 660e5e4c-0804-429b-ab18-9305a433391f] [ 611.961191] env[61987]: DEBUG nova.compute.utils [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Binding failed for port 5f60a42f-4af4-440f-93a1-79dacf25140a, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 611.963669] env[61987]: DEBUG nova.compute.manager [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Build of instance 660e5e4c-0804-429b-ab18-9305a433391f was re-scheduled: Binding failed for port 5f60a42f-4af4-440f-93a1-79dacf25140a, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 611.967283] env[61987]: DEBUG nova.compute.manager [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 611.967283] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Acquiring lock "refresh_cache-660e5e4c-0804-429b-ab18-9305a433391f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.967283] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Acquired lock "refresh_cache-660e5e4c-0804-429b-ab18-9305a433391f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.967283] env[61987]: DEBUG nova.network.neutron [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 611.967283] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.191s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.969208] env[61987]: INFO nova.compute.claims [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 611.989745] env[61987]: DEBUG nova.compute.manager [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 612.172671] env[61987]: INFO nova.compute.manager [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] [instance: bac5a3bc-5a58-4c1d-abb8-1aaf6151358a] Took 1.04 seconds to deallocate network for instance. [ 612.186356] env[61987]: DEBUG oslo_vmware.api [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Task: {'id': task-1061245, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.088895} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 612.186631] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 612.186810] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 612.187063] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 612.187179] env[61987]: INFO nova.compute.manager [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Took 1.08 seconds to destroy the instance on the hypervisor. [ 612.187467] env[61987]: DEBUG oslo.service.loopingcall [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 612.187548] env[61987]: DEBUG nova.compute.manager [-] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 612.187643] env[61987]: DEBUG nova.network.neutron [-] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 612.203797] env[61987]: DEBUG nova.network.neutron [-] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 612.502273] env[61987]: DEBUG nova.network.neutron [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 612.521826] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.671782] env[61987]: DEBUG nova.network.neutron [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.706420] env[61987]: DEBUG nova.network.neutron [-] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.175722] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Releasing lock "refresh_cache-660e5e4c-0804-429b-ab18-9305a433391f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.175967] env[61987]: DEBUG nova.compute.manager [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 613.176187] env[61987]: DEBUG nova.compute.manager [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 613.176480] env[61987]: DEBUG nova.network.neutron [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 613.211258] env[61987]: INFO nova.compute.manager [-] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Took 1.02 seconds to deallocate network for instance. [ 613.217731] env[61987]: INFO nova.scheduler.client.report [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Deleted allocations for instance bac5a3bc-5a58-4c1d-abb8-1aaf6151358a [ 613.381131] env[61987]: DEBUG nova.network.neutron [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 613.398863] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fbf5fde-d7cd-4493-99e9-00347c431e6f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.407137] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99f2adb3-7682-4450-a67a-2121baa08867 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.442863] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b44b66-a30b-40f7-ba02-e0d44ac19d37 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.450631] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9fd568-9d04-41b0-ac84-19111a523fad {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.468112] env[61987]: DEBUG nova.compute.provider_tree [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.724420] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.730274] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d8b0a340-0869-4f96-b2f4-622df1ce701e tempest-ServerActionsTestOtherB-1530843648 tempest-ServerActionsTestOtherB-1530843648-project-member] Lock "bac5a3bc-5a58-4c1d-abb8-1aaf6151358a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.838s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.887306] env[61987]: DEBUG nova.network.neutron [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.976227] env[61987]: DEBUG nova.scheduler.client.report [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 614.152424] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Acquiring lock "d55395cc-0b92-4e99-9b80-913c19f6c7f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.152746] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Lock "d55395cc-0b92-4e99-9b80-913c19f6c7f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.242025] env[61987]: DEBUG nova.compute.manager [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 614.393226] env[61987]: INFO nova.compute.manager [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: 660e5e4c-0804-429b-ab18-9305a433391f] Took 1.22 seconds to deallocate network for instance. [ 614.485024] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.515s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.485024] env[61987]: DEBUG nova.compute.manager [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 614.486359] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.872s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.767571] env[61987]: DEBUG oslo_concurrency.lockutils [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.994069] env[61987]: DEBUG nova.compute.utils [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 615.002148] env[61987]: DEBUG nova.compute.manager [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Not allocating networking since 'none' was specified. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 615.433059] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquiring lock "4edd0159-6db1-41bd-a616-31a34e984059" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.433059] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "4edd0159-6db1-41bd-a616-31a34e984059" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.436051] env[61987]: INFO nova.scheduler.client.report [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Deleted allocations for instance 660e5e4c-0804-429b-ab18-9305a433391f [ 615.459866] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6626b72b-a478-4880-b9c8-e198506f95a5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.473565] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4af93f6-bb66-40b2-93e1-edd099f9b99d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.511199] env[61987]: DEBUG nova.compute.manager [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 615.514556] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aefa959b-4713-4725-95f9-3a632a80166c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.527633] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02fa55ae-6fc4-458f-9ec8-1499c8cd6289 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.541049] env[61987]: DEBUG nova.compute.provider_tree [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.951811] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0194b169-313c-4eb0-95c4-3899b8e7510c tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Lock "660e5e4c-0804-429b-ab18-9305a433391f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.390s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.047104] env[61987]: DEBUG nova.scheduler.client.report [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 616.454851] env[61987]: DEBUG nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 616.523691] env[61987]: DEBUG nova.compute.manager [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 616.552801] env[61987]: DEBUG nova.virt.hardware [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 616.553186] env[61987]: DEBUG nova.virt.hardware [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 616.557045] env[61987]: DEBUG nova.virt.hardware [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 616.557045] env[61987]: DEBUG nova.virt.hardware [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 616.557045] env[61987]: DEBUG nova.virt.hardware [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 616.557045] env[61987]: DEBUG nova.virt.hardware [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 616.557045] env[61987]: DEBUG nova.virt.hardware [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 616.557255] env[61987]: DEBUG nova.virt.hardware [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 616.557255] env[61987]: DEBUG nova.virt.hardware [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 616.557255] env[61987]: DEBUG nova.virt.hardware [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 616.557255] env[61987]: DEBUG nova.virt.hardware [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 616.557255] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.069s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.557382] env[61987]: ERROR nova.compute.manager [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 45aa1dda-6ba4-4927-9269-11fba215aa5a, please check neutron logs for more information. [ 616.557382] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Traceback (most recent call last): [ 616.557382] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 616.557382] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] self.driver.spawn(context, instance, image_meta, [ 616.557382] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 616.557382] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] self._vmops.spawn(context, instance, image_meta, injected_files, [ 616.557382] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 616.557382] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] vm_ref = self.build_virtual_machine(instance, [ 616.557382] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 616.557382] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] vif_infos = vmwarevif.get_vif_info(self._session, [ 616.557382] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 616.557619] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] for vif in network_info: [ 616.557619] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 616.557619] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] return self._sync_wrapper(fn, *args, **kwargs) [ 616.557619] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 616.557619] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] self.wait() [ 616.557619] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 616.557619] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] self[:] = self._gt.wait() [ 616.557619] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 616.557619] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] return self._exit_event.wait() [ 616.557619] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 616.557619] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] current.throw(*self._exc) [ 616.557619] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.557619] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] result = function(*args, **kwargs) [ 616.557893] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 616.557893] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] return func(*args, **kwargs) [ 616.557893] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 616.557893] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] raise e [ 616.557893] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 616.557893] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] nwinfo = self.network_api.allocate_for_instance( [ 616.557893] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 616.557893] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] created_port_ids = self._update_ports_for_instance( [ 616.557893] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 616.557893] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] with excutils.save_and_reraise_exception(): [ 616.557893] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.557893] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] self.force_reraise() [ 616.557893] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.558188] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] raise self.value [ 616.558188] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 616.558188] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] updated_port = self._update_port( [ 616.558188] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.558188] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] _ensure_no_port_binding_failure(port) [ 616.558188] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.558188] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] raise exception.PortBindingFailed(port_id=port['id']) [ 616.558188] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] nova.exception.PortBindingFailed: Binding failed for port 45aa1dda-6ba4-4927-9269-11fba215aa5a, please check neutron logs for more information. [ 616.558188] env[61987]: ERROR nova.compute.manager [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] [ 616.558188] env[61987]: DEBUG nova.compute.utils [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Binding failed for port 45aa1dda-6ba4-4927-9269-11fba215aa5a, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 616.558593] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19422ec4-5c11-4b05-9c9d-ea40786c6f8d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.561573] env[61987]: DEBUG nova.compute.manager [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Build of instance c7bfab3b-bfb1-416f-a302-9956c4fb0748 was re-scheduled: Binding failed for port 45aa1dda-6ba4-4927-9269-11fba215aa5a, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 616.562227] env[61987]: DEBUG nova.compute.manager [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 616.562227] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Acquiring lock "refresh_cache-c7bfab3b-bfb1-416f-a302-9956c4fb0748" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.562378] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Acquired lock "refresh_cache-c7bfab3b-bfb1-416f-a302-9956c4fb0748" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.562540] env[61987]: DEBUG nova.network.neutron [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 616.563589] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.864s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.572967] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5597281a-66c7-47e4-ba44-f87b14e8833b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.586785] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Instance VIF info [] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 616.593328] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Creating folder: Project (80e97c469c7440a680687ebd5c17a935). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 616.593891] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f62c8ead-3962-43ea-8977-707c3c48bc83 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.605274] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Created folder: Project (80e97c469c7440a680687ebd5c17a935) in parent group-v234219. [ 616.605626] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Creating folder: Instances. Parent ref: group-v234223. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 616.605719] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-07ceebeb-cc93-482b-a4d0-0d57f1275147 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.614959] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Created folder: Instances in parent group-v234223. [ 616.615042] env[61987]: DEBUG oslo.service.loopingcall [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 616.615275] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 616.615484] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-344b442c-7044-41a3-8766-231b6f03e325 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.632189] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 616.632189] env[61987]: value = "task-1061248" [ 616.632189] env[61987]: _type = "Task" [ 616.632189] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.640735] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061248, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.976754] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.090638] env[61987]: DEBUG nova.network.neutron [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.142420] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061248, 'name': CreateVM_Task, 'duration_secs': 0.256687} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.144947] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 617.145832] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.145832] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.146153] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 617.146441] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-baaf7d66-6d10-4190-abc7-26774dcd4b78 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.152722] env[61987]: DEBUG oslo_vmware.api [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Waiting for the task: (returnval){ [ 617.152722] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]526f656f-30b6-6372-23e8-f7b7c32d2d9a" [ 617.152722] env[61987]: _type = "Task" [ 617.152722] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.163836] env[61987]: DEBUG oslo_vmware.api [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]526f656f-30b6-6372-23e8-f7b7c32d2d9a, 'name': SearchDatastore_Task, 'duration_secs': 0.008106} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.164178] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.164435] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 617.164678] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.164827] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.165008] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 617.165632] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9eaa6a6d-3cab-463e-b925-60bbbfe3ad80 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.172661] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 617.172661] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 617.172931] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2da298c-7061-4a30-bdea-6f3111337aef {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.177957] env[61987]: DEBUG oslo_vmware.api [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Waiting for the task: (returnval){ [ 617.177957] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52426e35-62db-af14-4160-79a6e4bb1217" [ 617.177957] env[61987]: _type = "Task" [ 617.177957] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.188764] env[61987]: DEBUG oslo_vmware.api [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52426e35-62db-af14-4160-79a6e4bb1217, 'name': SearchDatastore_Task, 'duration_secs': 0.008198} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.189618] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e14a444-ecee-4884-9589-74982bf529ed {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.194397] env[61987]: DEBUG oslo_vmware.api [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Waiting for the task: (returnval){ [ 617.194397] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]528e005f-9216-e680-3534-3db8ae4da20d" [ 617.194397] env[61987]: _type = "Task" [ 617.194397] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.204386] env[61987]: DEBUG oslo_vmware.api [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]528e005f-9216-e680-3534-3db8ae4da20d, 'name': SearchDatastore_Task, 'duration_secs': 0.007367} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.204632] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.204882] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] e6c56d86-e428-49fb-9944-293ad5e5948c/e6c56d86-e428-49fb-9944-293ad5e5948c.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 617.205144] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-74d1618c-9bb2-4f49-9e5a-de8559dab5d0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.211294] env[61987]: DEBUG oslo_vmware.api [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Waiting for the task: (returnval){ [ 617.211294] env[61987]: value = "task-1061249" [ 617.211294] env[61987]: _type = "Task" [ 617.211294] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.221278] env[61987]: DEBUG oslo_vmware.api [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Task: {'id': task-1061249, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.222199] env[61987]: DEBUG nova.network.neutron [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.544958] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62cae0a8-ec9c-42ce-9c3e-c3ccc2cb88e8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.554904] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc7510b-ee06-464a-bdd6-aae28fbf8ab1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.593487] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c25b36c3-3b39-492e-97bd-885f486fc4d7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.602972] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f7cdb0-ef42-4219-a9d0-5478ae8df7df {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.619348] env[61987]: DEBUG nova.compute.provider_tree [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 617.723602] env[61987]: DEBUG oslo_vmware.api [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Task: {'id': task-1061249, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.466938} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 617.724119] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Releasing lock "refresh_cache-c7bfab3b-bfb1-416f-a302-9956c4fb0748" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.724353] env[61987]: DEBUG nova.compute.manager [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 617.724535] env[61987]: DEBUG nova.compute.manager [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 617.724705] env[61987]: DEBUG nova.network.neutron [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 617.726410] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] e6c56d86-e428-49fb-9944-293ad5e5948c/e6c56d86-e428-49fb-9944-293ad5e5948c.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 617.726629] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 617.726888] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-00cd4176-ade9-476f-90c7-b7d237e8be67 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.733118] env[61987]: DEBUG oslo_vmware.api [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Waiting for the task: (returnval){ [ 617.733118] env[61987]: value = "task-1061250" [ 617.733118] env[61987]: _type = "Task" [ 617.733118] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.743499] env[61987]: DEBUG oslo_vmware.api [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Task: {'id': task-1061250, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.893621] env[61987]: DEBUG nova.network.neutron [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 618.126185] env[61987]: DEBUG nova.scheduler.client.report [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 618.245869] env[61987]: DEBUG oslo_vmware.api [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Task: {'id': task-1061250, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061517} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.246215] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 618.247050] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee800cc-6b95-467b-baf8-34e411c4ceee {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.267611] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Reconfiguring VM instance instance-00000015 to attach disk [datastore2] e6c56d86-e428-49fb-9944-293ad5e5948c/e6c56d86-e428-49fb-9944-293ad5e5948c.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 618.267973] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-216b576e-65f3-4bb0-9749-0872798e0ce8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.287879] env[61987]: DEBUG oslo_vmware.api [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Waiting for the task: (returnval){ [ 618.287879] env[61987]: value = "task-1061251" [ 618.287879] env[61987]: _type = "Task" [ 618.287879] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.297111] env[61987]: DEBUG oslo_vmware.api [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Task: {'id': task-1061251, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.397011] env[61987]: DEBUG nova.network.neutron [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.629601] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.066s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.630259] env[61987]: ERROR nova.compute.manager [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 36e53101-e4f1-4e24-b454-a7cda6c93d8d, please check neutron logs for more information. [ 618.630259] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Traceback (most recent call last): [ 618.630259] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 618.630259] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] self.driver.spawn(context, instance, image_meta, [ 618.630259] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 618.630259] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.630259] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.630259] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] vm_ref = self.build_virtual_machine(instance, [ 618.630259] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.630259] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.630259] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.630524] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] for vif in network_info: [ 618.630524] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.630524] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] return self._sync_wrapper(fn, *args, **kwargs) [ 618.630524] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.630524] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] self.wait() [ 618.630524] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.630524] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] self[:] = self._gt.wait() [ 618.630524] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.630524] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] return self._exit_event.wait() [ 618.630524] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 618.630524] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] current.throw(*self._exc) [ 618.630524] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.630524] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] result = function(*args, **kwargs) [ 618.630800] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.630800] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] return func(*args, **kwargs) [ 618.630800] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 618.630800] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] raise e [ 618.630800] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 618.630800] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] nwinfo = self.network_api.allocate_for_instance( [ 618.630800] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 618.630800] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] created_port_ids = self._update_ports_for_instance( [ 618.630800] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 618.630800] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] with excutils.save_and_reraise_exception(): [ 618.630800] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.630800] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] self.force_reraise() [ 618.630800] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.631151] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] raise self.value [ 618.631151] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 618.631151] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] updated_port = self._update_port( [ 618.631151] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.631151] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] _ensure_no_port_binding_failure(port) [ 618.631151] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.631151] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] raise exception.PortBindingFailed(port_id=port['id']) [ 618.631151] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] nova.exception.PortBindingFailed: Binding failed for port 36e53101-e4f1-4e24-b454-a7cda6c93d8d, please check neutron logs for more information. [ 618.631151] env[61987]: ERROR nova.compute.manager [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] [ 618.631151] env[61987]: DEBUG nova.compute.utils [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Binding failed for port 36e53101-e4f1-4e24-b454-a7cda6c93d8d, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 618.632780] env[61987]: DEBUG nova.compute.manager [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Build of instance 1915d6b8-2798-4e35-9a35-4191659c2cb6 was re-scheduled: Binding failed for port 36e53101-e4f1-4e24-b454-a7cda6c93d8d, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 618.633247] env[61987]: DEBUG nova.compute.manager [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 618.633478] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "refresh_cache-1915d6b8-2798-4e35-9a35-4191659c2cb6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.633627] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired lock "refresh_cache-1915d6b8-2798-4e35-9a35-4191659c2cb6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.633784] env[61987]: DEBUG nova.network.neutron [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 618.634769] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.364s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.636266] env[61987]: INFO nova.compute.claims [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 618.801432] env[61987]: DEBUG oslo_vmware.api [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Task: {'id': task-1061251, 'name': ReconfigVM_Task, 'duration_secs': 0.263945} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.803342] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Reconfigured VM instance instance-00000015 to attach disk [datastore2] e6c56d86-e428-49fb-9944-293ad5e5948c/e6c56d86-e428-49fb-9944-293ad5e5948c.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 618.803342] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-841c7191-ccaa-454f-b81b-dbeed63db7c9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.811053] env[61987]: DEBUG oslo_vmware.api [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Waiting for the task: (returnval){ [ 618.811053] env[61987]: value = "task-1061252" [ 618.811053] env[61987]: _type = "Task" [ 618.811053] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.820546] env[61987]: DEBUG oslo_vmware.api [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Task: {'id': task-1061252, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.902702] env[61987]: INFO nova.compute.manager [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] [instance: c7bfab3b-bfb1-416f-a302-9956c4fb0748] Took 1.18 seconds to deallocate network for instance. [ 619.173987] env[61987]: DEBUG nova.network.neutron [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.321759] env[61987]: DEBUG oslo_vmware.api [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Task: {'id': task-1061252, 'name': Rename_Task, 'duration_secs': 0.142614} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.322011] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 619.326378] env[61987]: DEBUG nova.network.neutron [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.327562] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-817736c2-9471-4108-a0c5-44b9e6a5c21e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.337770] env[61987]: DEBUG oslo_vmware.api [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Waiting for the task: (returnval){ [ 619.337770] env[61987]: value = "task-1061253" [ 619.337770] env[61987]: _type = "Task" [ 619.337770] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.346516] env[61987]: DEBUG oslo_vmware.api [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Task: {'id': task-1061253, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.833573] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Releasing lock "refresh_cache-1915d6b8-2798-4e35-9a35-4191659c2cb6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.833727] env[61987]: DEBUG nova.compute.manager [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 619.833938] env[61987]: DEBUG nova.compute.manager [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 619.834241] env[61987]: DEBUG nova.network.neutron [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 619.848280] env[61987]: DEBUG oslo_vmware.api [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Task: {'id': task-1061253, 'name': PowerOnVM_Task, 'duration_secs': 0.401008} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.848542] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 619.848812] env[61987]: INFO nova.compute.manager [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Took 3.32 seconds to spawn the instance on the hypervisor. [ 619.849023] env[61987]: DEBUG nova.compute.manager [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 619.851874] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9a1218-7bc6-460f-b472-821e970a8828 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.876136] env[61987]: DEBUG nova.network.neutron [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.929068] env[61987]: INFO nova.scheduler.client.report [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Deleted allocations for instance c7bfab3b-bfb1-416f-a302-9956c4fb0748 [ 620.162375] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0509704e-cd69-4f3b-92ab-24bf3b5d42a6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.171060] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd7336c-7f27-4beb-aa85-1f524bb4e41b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.203985] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6950a7cd-60a7-4c34-89f6-8c7fbc40adb7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.213150] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-201c9579-4a5e-40f3-b64d-2da40cf0672f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.229777] env[61987]: DEBUG nova.compute.provider_tree [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.378941] env[61987]: DEBUG nova.network.neutron [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.381348] env[61987]: INFO nova.compute.manager [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Took 21.62 seconds to build instance. [ 620.440048] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b40ee998-44f1-4fb4-b7cb-bb266fbd5277 tempest-ServersAdminTestJSON-444365908 tempest-ServersAdminTestJSON-444365908-project-member] Lock "c7bfab3b-bfb1-416f-a302-9956c4fb0748" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.403s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.733198] env[61987]: DEBUG nova.scheduler.client.report [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 620.883244] env[61987]: INFO nova.compute.manager [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 1915d6b8-2798-4e35-9a35-4191659c2cb6] Took 1.05 seconds to deallocate network for instance. [ 620.889382] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c39cd4a-dd17-4360-89ab-cfbd31bcd500 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Lock "e6c56d86-e428-49fb-9944-293ad5e5948c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.688s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.946280] env[61987]: DEBUG nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 621.243022] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.607s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.243855] env[61987]: DEBUG nova.compute.manager [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 621.248464] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.984s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.251113] env[61987]: INFO nova.compute.claims [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 621.398503] env[61987]: DEBUG nova.compute.manager [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 621.477606] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.760496] env[61987]: DEBUG nova.compute.utils [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 621.764372] env[61987]: DEBUG nova.compute.manager [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 621.764613] env[61987]: DEBUG nova.network.neutron [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 621.877463] env[61987]: DEBUG nova.policy [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4ead7955f1d475590523ac98c579956', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '315a2a3450024ed38e27cd7758f392f9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 621.925963] env[61987]: INFO nova.scheduler.client.report [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Deleted allocations for instance 1915d6b8-2798-4e35-9a35-4191659c2cb6 [ 621.938309] env[61987]: DEBUG nova.compute.manager [None req-c22f7c2b-20a7-47d5-8821-e4e61a8108df tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 621.939132] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b8d68b-ae61-41c6-ba57-67ab9cf74a1a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.943128] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.101019] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Acquiring lock "e6c56d86-e428-49fb-9944-293ad5e5948c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.101019] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Lock "e6c56d86-e428-49fb-9944-293ad5e5948c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.101019] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Acquiring lock "e6c56d86-e428-49fb-9944-293ad5e5948c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.101019] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Lock "e6c56d86-e428-49fb-9944-293ad5e5948c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.101239] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Lock "e6c56d86-e428-49fb-9944-293ad5e5948c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.103286] env[61987]: INFO nova.compute.manager [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Terminating instance [ 622.265049] env[61987]: DEBUG nova.compute.manager [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 622.445168] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ccf56f8f-ed7d-4892-b914-ef21f11ee353 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "1915d6b8-2798-4e35-9a35-4191659c2cb6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.494s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.450605] env[61987]: INFO nova.compute.manager [None req-c22f7c2b-20a7-47d5-8821-e4e61a8108df tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] instance snapshotting [ 622.451402] env[61987]: DEBUG nova.objects.instance [None req-c22f7c2b-20a7-47d5-8821-e4e61a8108df tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Lazy-loading 'flavor' on Instance uuid e6c56d86-e428-49fb-9944-293ad5e5948c {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 622.509442] env[61987]: DEBUG nova.network.neutron [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Successfully created port: cf948a39-baf8-4fe5-b691-cea4f5698e8f {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 622.607131] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Acquiring lock "refresh_cache-e6c56d86-e428-49fb-9944-293ad5e5948c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.607336] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Acquired lock "refresh_cache-e6c56d86-e428-49fb-9944-293ad5e5948c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.607516] env[61987]: DEBUG nova.network.neutron [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 622.751253] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8f8e24d-2385-46e3-9f87-ae3b38419a5c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.763393] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2196be89-11b5-4634-9398-b182390ecba0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.799624] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00b5a513-f447-4232-9bc9-b275202154d0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.808846] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aeb637d-472c-4a55-b1d2-1ae801ed1025 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.825162] env[61987]: DEBUG nova.compute.provider_tree [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.951657] env[61987]: DEBUG nova.compute.manager [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 622.964221] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43d3c2ce-c12f-4c83-bb2c-44388a6d1a9e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.984553] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d78d1a2-7cce-47db-a646-27ffe04805f5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.131858] env[61987]: DEBUG nova.network.neutron [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 623.236217] env[61987]: DEBUG nova.network.neutron [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.301833] env[61987]: DEBUG nova.compute.manager [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 623.331023] env[61987]: DEBUG nova.scheduler.client.report [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 623.356090] env[61987]: DEBUG nova.virt.hardware [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 623.356090] env[61987]: DEBUG nova.virt.hardware [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 623.356090] env[61987]: DEBUG nova.virt.hardware [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 623.356391] env[61987]: DEBUG nova.virt.hardware [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 623.356391] env[61987]: DEBUG nova.virt.hardware [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 623.356391] env[61987]: DEBUG nova.virt.hardware [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 623.356391] env[61987]: DEBUG nova.virt.hardware [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 623.356391] env[61987]: DEBUG nova.virt.hardware [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 623.356569] env[61987]: DEBUG nova.virt.hardware [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 623.356569] env[61987]: DEBUG nova.virt.hardware [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 623.356569] env[61987]: DEBUG nova.virt.hardware [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 623.357519] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71bef147-a240-437c-ad54-858671076c2e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.368054] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d464c28f-1104-445d-99b9-f87dbad2e70d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.491408] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.495551] env[61987]: DEBUG nova.compute.manager [None req-c22f7c2b-20a7-47d5-8821-e4e61a8108df tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Instance disappeared during snapshot {{(pid=61987) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4594}} [ 623.591404] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 623.591499] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 623.670230] env[61987]: DEBUG nova.compute.manager [None req-c22f7c2b-20a7-47d5-8821-e4e61a8108df tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Found 0 images (rotation: 2) {{(pid=61987) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5016}} [ 623.756831] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Releasing lock "refresh_cache-e6c56d86-e428-49fb-9944-293ad5e5948c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.756831] env[61987]: DEBUG nova.compute.manager [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 623.756831] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 623.756831] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fa6d505-2a8a-45b8-9dca-81f50422b950 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.765652] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 623.765812] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0dc7a319-8bb1-4490-bf31-5328b97582cc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.775171] env[61987]: DEBUG oslo_vmware.api [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Waiting for the task: (returnval){ [ 623.775171] env[61987]: value = "task-1061258" [ 623.775171] env[61987]: _type = "Task" [ 623.775171] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.782813] env[61987]: DEBUG oslo_vmware.api [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Task: {'id': task-1061258, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.843645] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.595s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.844297] env[61987]: DEBUG nova.compute.manager [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 623.847866] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.630s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.849297] env[61987]: INFO nova.compute.claims [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 624.103356] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 624.103356] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Starting heal instance info cache {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10401}} [ 624.103356] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Rebuilding the list of instances to heal {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10405}} [ 624.284856] env[61987]: DEBUG oslo_vmware.api [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Task: {'id': task-1061258, 'name': PowerOffVM_Task, 'duration_secs': 0.168347} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.285177] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 624.285340] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 624.285590] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-18571651-862d-4c3b-b8ce-594209309d2c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.314726] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 624.314726] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 624.314726] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Deleting the datastore file [datastore2] e6c56d86-e428-49fb-9944-293ad5e5948c {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 624.315752] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-11a7001d-b3be-4b43-8a02-c8ad49ce5b0d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.319064] env[61987]: DEBUG nova.compute.manager [req-ac29f06b-fd22-439c-8752-afcb0432f2e2 req-446fa48b-dc8d-4a2d-8cdd-55378d32e393 service nova] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Received event network-changed-cf948a39-baf8-4fe5-b691-cea4f5698e8f {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 624.319292] env[61987]: DEBUG nova.compute.manager [req-ac29f06b-fd22-439c-8752-afcb0432f2e2 req-446fa48b-dc8d-4a2d-8cdd-55378d32e393 service nova] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Refreshing instance network info cache due to event network-changed-cf948a39-baf8-4fe5-b691-cea4f5698e8f. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 624.319538] env[61987]: DEBUG oslo_concurrency.lockutils [req-ac29f06b-fd22-439c-8752-afcb0432f2e2 req-446fa48b-dc8d-4a2d-8cdd-55378d32e393 service nova] Acquiring lock "refresh_cache-40fefd68-3d6f-4451-997f-e8fa49847dea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.319721] env[61987]: DEBUG oslo_concurrency.lockutils [req-ac29f06b-fd22-439c-8752-afcb0432f2e2 req-446fa48b-dc8d-4a2d-8cdd-55378d32e393 service nova] Acquired lock "refresh_cache-40fefd68-3d6f-4451-997f-e8fa49847dea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.319968] env[61987]: DEBUG nova.network.neutron [req-ac29f06b-fd22-439c-8752-afcb0432f2e2 req-446fa48b-dc8d-4a2d-8cdd-55378d32e393 service nova] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Refreshing network info cache for port cf948a39-baf8-4fe5-b691-cea4f5698e8f {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 624.330434] env[61987]: DEBUG oslo_vmware.api [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Waiting for the task: (returnval){ [ 624.330434] env[61987]: value = "task-1061260" [ 624.330434] env[61987]: _type = "Task" [ 624.330434] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.341838] env[61987]: DEBUG oslo_vmware.api [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Task: {'id': task-1061260, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.358575] env[61987]: DEBUG nova.compute.utils [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 624.363106] env[61987]: ERROR nova.compute.manager [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cf948a39-baf8-4fe5-b691-cea4f5698e8f, please check neutron logs for more information. [ 624.363106] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 624.363106] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 624.363106] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 624.363106] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 624.363106] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 624.363106] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 624.363106] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 624.363106] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.363106] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 624.363106] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.363106] env[61987]: ERROR nova.compute.manager raise self.value [ 624.363106] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 624.363106] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 624.363106] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.363106] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 624.363835] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.363835] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 624.363835] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cf948a39-baf8-4fe5-b691-cea4f5698e8f, please check neutron logs for more information. [ 624.363835] env[61987]: ERROR nova.compute.manager [ 624.363835] env[61987]: Traceback (most recent call last): [ 624.363835] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 624.363835] env[61987]: listener.cb(fileno) [ 624.363835] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.363835] env[61987]: result = function(*args, **kwargs) [ 624.363835] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 624.363835] env[61987]: return func(*args, **kwargs) [ 624.363835] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 624.363835] env[61987]: raise e [ 624.363835] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 624.363835] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 624.363835] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 624.363835] env[61987]: created_port_ids = self._update_ports_for_instance( [ 624.363835] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 624.363835] env[61987]: with excutils.save_and_reraise_exception(): [ 624.363835] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.363835] env[61987]: self.force_reraise() [ 624.363835] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.363835] env[61987]: raise self.value [ 624.363835] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 624.363835] env[61987]: updated_port = self._update_port( [ 624.363835] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.363835] env[61987]: _ensure_no_port_binding_failure(port) [ 624.363835] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.363835] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 624.366054] env[61987]: nova.exception.PortBindingFailed: Binding failed for port cf948a39-baf8-4fe5-b691-cea4f5698e8f, please check neutron logs for more information. [ 624.366054] env[61987]: Removing descriptor: 18 [ 624.366054] env[61987]: DEBUG nova.compute.manager [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 624.366054] env[61987]: DEBUG nova.network.neutron [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 624.366853] env[61987]: ERROR nova.compute.manager [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cf948a39-baf8-4fe5-b691-cea4f5698e8f, please check neutron logs for more information. [ 624.366853] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Traceback (most recent call last): [ 624.366853] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 624.366853] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] yield resources [ 624.366853] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 624.366853] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] self.driver.spawn(context, instance, image_meta, [ 624.366853] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 624.366853] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.366853] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.366853] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] vm_ref = self.build_virtual_machine(instance, [ 624.366853] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.367523] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.367523] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.367523] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] for vif in network_info: [ 624.367523] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 624.367523] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] return self._sync_wrapper(fn, *args, **kwargs) [ 624.367523] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 624.367523] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] self.wait() [ 624.367523] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 624.367523] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] self[:] = self._gt.wait() [ 624.367523] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.367523] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] return self._exit_event.wait() [ 624.367523] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.367523] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] result = hub.switch() [ 624.368079] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.368079] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] return self.greenlet.switch() [ 624.368079] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.368079] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] result = function(*args, **kwargs) [ 624.368079] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 624.368079] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] return func(*args, **kwargs) [ 624.368079] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 624.368079] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] raise e [ 624.368079] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 624.368079] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] nwinfo = self.network_api.allocate_for_instance( [ 624.368079] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 624.368079] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] created_port_ids = self._update_ports_for_instance( [ 624.368079] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 624.369146] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] with excutils.save_and_reraise_exception(): [ 624.369146] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.369146] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] self.force_reraise() [ 624.369146] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.369146] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] raise self.value [ 624.369146] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 624.369146] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] updated_port = self._update_port( [ 624.369146] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.369146] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] _ensure_no_port_binding_failure(port) [ 624.369146] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.369146] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] raise exception.PortBindingFailed(port_id=port['id']) [ 624.369146] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] nova.exception.PortBindingFailed: Binding failed for port cf948a39-baf8-4fe5-b691-cea4f5698e8f, please check neutron logs for more information. [ 624.369146] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] [ 624.369761] env[61987]: INFO nova.compute.manager [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Terminating instance [ 624.454630] env[61987]: DEBUG nova.policy [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '27ae5b57556245e7b27ccc1363485352', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a53eb8ad8274fe18fd149c44eb0ffbb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 624.606245] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Skipping network cache update for instance because it is being deleted. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10418}} [ 624.606245] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Skipping network cache update for instance because it is Building. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10414}} [ 624.606245] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Skipping network cache update for instance because it is Building. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10414}} [ 624.606499] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Skipping network cache update for instance because it is Building. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10414}} [ 624.646025] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "refresh_cache-6bf0d1f1-d079-4956-bc9a-86ba66885f3a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.646202] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquired lock "refresh_cache-6bf0d1f1-d079-4956-bc9a-86ba66885f3a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.646357] env[61987]: DEBUG nova.network.neutron [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Forcefully refreshing network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 624.646569] env[61987]: DEBUG nova.objects.instance [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lazy-loading 'info_cache' on Instance uuid 6bf0d1f1-d079-4956-bc9a-86ba66885f3a {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 624.821192] env[61987]: DEBUG oslo_concurrency.lockutils [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "325183b9-e35b-4f9c-a271-d0fdb05288bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.821192] env[61987]: DEBUG oslo_concurrency.lockutils [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "325183b9-e35b-4f9c-a271-d0fdb05288bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.844126] env[61987]: DEBUG oslo_vmware.api [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Task: {'id': task-1061260, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.086311} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.844531] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 624.844837] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 624.845160] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 624.847782] env[61987]: INFO nova.compute.manager [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Took 1.09 seconds to destroy the instance on the hypervisor. [ 624.847782] env[61987]: DEBUG oslo.service.loopingcall [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 624.847782] env[61987]: DEBUG nova.compute.manager [-] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 624.847782] env[61987]: DEBUG nova.network.neutron [-] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 624.875738] env[61987]: DEBUG nova.compute.manager [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 624.876467] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "refresh_cache-40fefd68-3d6f-4451-997f-e8fa49847dea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.908919] env[61987]: DEBUG nova.network.neutron [-] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 624.913147] env[61987]: DEBUG nova.network.neutron [req-ac29f06b-fd22-439c-8752-afcb0432f2e2 req-446fa48b-dc8d-4a2d-8cdd-55378d32e393 service nova] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.141018] env[61987]: DEBUG nova.network.neutron [req-ac29f06b-fd22-439c-8752-afcb0432f2e2 req-446fa48b-dc8d-4a2d-8cdd-55378d32e393 service nova] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.288871] env[61987]: DEBUG nova.network.neutron [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Successfully created port: 1c4aa77c-6462-4435-bda8-10dc64c8e052 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 625.385561] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4c10dbc-aeb1-4d6c-92e2-ce1e043cd50d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.393996] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ddd9c48-81bb-4c99-b12b-248d0eb7895d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.433032] env[61987]: DEBUG nova.network.neutron [-] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.435379] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-105cf8a8-1b01-47e4-8432-5fb6727f97a0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.444680] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8e61e1f-c213-4387-b468-eb025ea68608 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.459881] env[61987]: DEBUG nova.compute.provider_tree [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.643985] env[61987]: DEBUG oslo_concurrency.lockutils [req-ac29f06b-fd22-439c-8752-afcb0432f2e2 req-446fa48b-dc8d-4a2d-8cdd-55378d32e393 service nova] Releasing lock "refresh_cache-40fefd68-3d6f-4451-997f-e8fa49847dea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.644603] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired lock "refresh_cache-40fefd68-3d6f-4451-997f-e8fa49847dea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.644603] env[61987]: DEBUG nova.network.neutron [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 625.687292] env[61987]: DEBUG nova.network.neutron [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.892021] env[61987]: DEBUG nova.compute.manager [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 625.919386] env[61987]: DEBUG nova.virt.hardware [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 625.919742] env[61987]: DEBUG nova.virt.hardware [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 625.920037] env[61987]: DEBUG nova.virt.hardware [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 625.920796] env[61987]: DEBUG nova.virt.hardware [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 625.920939] env[61987]: DEBUG nova.virt.hardware [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 625.921203] env[61987]: DEBUG nova.virt.hardware [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 625.921472] env[61987]: DEBUG nova.virt.hardware [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 625.921671] env[61987]: DEBUG nova.virt.hardware [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 625.921871] env[61987]: DEBUG nova.virt.hardware [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 625.922085] env[61987]: DEBUG nova.virt.hardware [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 625.922432] env[61987]: DEBUG nova.virt.hardware [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 625.923397] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d9b5105-970a-455a-ad1b-0a53411cb67a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.932140] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a6fd43-ce31-4523-9d26-09612c7706f6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.936565] env[61987]: INFO nova.compute.manager [-] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Took 1.09 seconds to deallocate network for instance. [ 625.963303] env[61987]: DEBUG nova.scheduler.client.report [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 626.175419] env[61987]: DEBUG nova.network.neutron [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.320448] env[61987]: DEBUG nova.network.neutron [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.352833] env[61987]: DEBUG nova.network.neutron [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.357927] env[61987]: DEBUG nova.compute.manager [req-3460a19b-a826-4f1c-b320-345d57d19eef req-12191976-c1c2-4e1a-bf1d-045b89c436e6 service nova] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Received event network-vif-deleted-cf948a39-baf8-4fe5-b691-cea4f5698e8f {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 626.358267] env[61987]: DEBUG nova.compute.manager [req-3460a19b-a826-4f1c-b320-345d57d19eef req-12191976-c1c2-4e1a-bf1d-045b89c436e6 service nova] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Received event network-changed-1c4aa77c-6462-4435-bda8-10dc64c8e052 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 626.358654] env[61987]: DEBUG nova.compute.manager [req-3460a19b-a826-4f1c-b320-345d57d19eef req-12191976-c1c2-4e1a-bf1d-045b89c436e6 service nova] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Refreshing instance network info cache due to event network-changed-1c4aa77c-6462-4435-bda8-10dc64c8e052. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 626.359228] env[61987]: DEBUG oslo_concurrency.lockutils [req-3460a19b-a826-4f1c-b320-345d57d19eef req-12191976-c1c2-4e1a-bf1d-045b89c436e6 service nova] Acquiring lock "refresh_cache-c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.359493] env[61987]: DEBUG oslo_concurrency.lockutils [req-3460a19b-a826-4f1c-b320-345d57d19eef req-12191976-c1c2-4e1a-bf1d-045b89c436e6 service nova] Acquired lock "refresh_cache-c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.359751] env[61987]: DEBUG nova.network.neutron [req-3460a19b-a826-4f1c-b320-345d57d19eef req-12191976-c1c2-4e1a-bf1d-045b89c436e6 service nova] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Refreshing network info cache for port 1c4aa77c-6462-4435-bda8-10dc64c8e052 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 626.453461] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.471540] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.621s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.471540] env[61987]: DEBUG nova.compute.manager [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 626.477900] env[61987]: DEBUG oslo_concurrency.lockutils [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.003s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.480130] env[61987]: INFO nova.compute.claims [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 626.568676] env[61987]: ERROR nova.compute.manager [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1c4aa77c-6462-4435-bda8-10dc64c8e052, please check neutron logs for more information. [ 626.568676] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 626.568676] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 626.568676] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 626.568676] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 626.568676] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 626.568676] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 626.568676] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 626.568676] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.568676] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 626.568676] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.568676] env[61987]: ERROR nova.compute.manager raise self.value [ 626.568676] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 626.568676] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 626.568676] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.568676] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 626.569218] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.569218] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 626.569218] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1c4aa77c-6462-4435-bda8-10dc64c8e052, please check neutron logs for more information. [ 626.569218] env[61987]: ERROR nova.compute.manager [ 626.569218] env[61987]: Traceback (most recent call last): [ 626.569218] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 626.569218] env[61987]: listener.cb(fileno) [ 626.569218] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.569218] env[61987]: result = function(*args, **kwargs) [ 626.569218] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 626.569218] env[61987]: return func(*args, **kwargs) [ 626.569218] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 626.569218] env[61987]: raise e [ 626.569218] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 626.569218] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 626.569218] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 626.569218] env[61987]: created_port_ids = self._update_ports_for_instance( [ 626.569218] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 626.569218] env[61987]: with excutils.save_and_reraise_exception(): [ 626.569218] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.569218] env[61987]: self.force_reraise() [ 626.569218] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.569218] env[61987]: raise self.value [ 626.569218] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 626.569218] env[61987]: updated_port = self._update_port( [ 626.569218] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.569218] env[61987]: _ensure_no_port_binding_failure(port) [ 626.569218] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.569218] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 626.569963] env[61987]: nova.exception.PortBindingFailed: Binding failed for port 1c4aa77c-6462-4435-bda8-10dc64c8e052, please check neutron logs for more information. [ 626.569963] env[61987]: Removing descriptor: 18 [ 626.569963] env[61987]: ERROR nova.compute.manager [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1c4aa77c-6462-4435-bda8-10dc64c8e052, please check neutron logs for more information. [ 626.569963] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Traceback (most recent call last): [ 626.569963] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 626.569963] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] yield resources [ 626.569963] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 626.569963] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] self.driver.spawn(context, instance, image_meta, [ 626.569963] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 626.569963] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] self._vmops.spawn(context, instance, image_meta, injected_files, [ 626.569963] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 626.569963] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] vm_ref = self.build_virtual_machine(instance, [ 626.570293] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 626.570293] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] vif_infos = vmwarevif.get_vif_info(self._session, [ 626.570293] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 626.570293] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] for vif in network_info: [ 626.570293] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 626.570293] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] return self._sync_wrapper(fn, *args, **kwargs) [ 626.570293] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 626.570293] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] self.wait() [ 626.570293] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 626.570293] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] self[:] = self._gt.wait() [ 626.570293] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 626.570293] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] return self._exit_event.wait() [ 626.570293] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 626.570759] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] result = hub.switch() [ 626.570759] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 626.570759] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] return self.greenlet.switch() [ 626.570759] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.570759] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] result = function(*args, **kwargs) [ 626.570759] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 626.570759] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] return func(*args, **kwargs) [ 626.570759] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 626.570759] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] raise e [ 626.570759] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 626.570759] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] nwinfo = self.network_api.allocate_for_instance( [ 626.570759] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 626.570759] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] created_port_ids = self._update_ports_for_instance( [ 626.571124] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 626.571124] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] with excutils.save_and_reraise_exception(): [ 626.571124] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.571124] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] self.force_reraise() [ 626.571124] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.571124] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] raise self.value [ 626.571124] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 626.571124] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] updated_port = self._update_port( [ 626.571124] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.571124] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] _ensure_no_port_binding_failure(port) [ 626.571124] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.571124] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] raise exception.PortBindingFailed(port_id=port['id']) [ 626.571453] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] nova.exception.PortBindingFailed: Binding failed for port 1c4aa77c-6462-4435-bda8-10dc64c8e052, please check neutron logs for more information. [ 626.571453] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] [ 626.571453] env[61987]: INFO nova.compute.manager [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Terminating instance [ 626.824681] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Releasing lock "refresh_cache-6bf0d1f1-d079-4956-bc9a-86ba66885f3a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.825311] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Updated the network info_cache for instance {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10472}} [ 626.825581] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 626.825951] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 626.826146] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 626.826312] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 626.826458] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 626.827687] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 626.827687] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61987) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11020}} [ 626.827687] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 626.866611] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Releasing lock "refresh_cache-40fefd68-3d6f-4451-997f-e8fa49847dea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.868327] env[61987]: DEBUG nova.compute.manager [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 626.868659] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 626.871253] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7c4f4f22-917b-4fcd-9187-428091a589dd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.879993] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4475905e-e4d3-47f7-8ae7-272c8a06a24a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.893690] env[61987]: DEBUG nova.network.neutron [req-3460a19b-a826-4f1c-b320-345d57d19eef req-12191976-c1c2-4e1a-bf1d-045b89c436e6 service nova] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.909932] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 40fefd68-3d6f-4451-997f-e8fa49847dea could not be found. [ 626.909932] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 626.909932] env[61987]: INFO nova.compute.manager [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Took 0.04 seconds to destroy the instance on the hypervisor. [ 626.909932] env[61987]: DEBUG oslo.service.loopingcall [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 626.909932] env[61987]: DEBUG nova.compute.manager [-] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 626.909932] env[61987]: DEBUG nova.network.neutron [-] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 626.939989] env[61987]: DEBUG nova.network.neutron [-] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 626.991097] env[61987]: DEBUG nova.compute.utils [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 626.994102] env[61987]: DEBUG nova.compute.manager [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 626.994102] env[61987]: DEBUG nova.network.neutron [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 627.030737] env[61987]: DEBUG nova.network.neutron [req-3460a19b-a826-4f1c-b320-345d57d19eef req-12191976-c1c2-4e1a-bf1d-045b89c436e6 service nova] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.064972] env[61987]: DEBUG nova.policy [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '87a8bf3961ce4d0c8e6afec8d26e8c04', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7d8c827ae5be4898a21417c0084d66dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 627.077335] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Acquiring lock "refresh_cache-c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.332989] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.410773] env[61987]: DEBUG nova.network.neutron [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Successfully created port: de0f490b-f7ab-4dfe-83b1-3574487c4556 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 627.442388] env[61987]: DEBUG nova.network.neutron [-] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.497283] env[61987]: DEBUG nova.compute.manager [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 627.540696] env[61987]: DEBUG oslo_concurrency.lockutils [req-3460a19b-a826-4f1c-b320-345d57d19eef req-12191976-c1c2-4e1a-bf1d-045b89c436e6 service nova] Releasing lock "refresh_cache-c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.541122] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Acquired lock "refresh_cache-c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.541320] env[61987]: DEBUG nova.network.neutron [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 627.947503] env[61987]: INFO nova.compute.manager [-] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Took 1.04 seconds to deallocate network for instance. [ 627.949405] env[61987]: DEBUG nova.compute.claims [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 627.949405] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.971355] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c7e4fb5-eb00-4986-8b4d-b2756f62179d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.982015] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37083ae4-3c84-4f63-acf5-b414ba57719a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.020019] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4583a7f-f34b-41a9-8200-134b6495c387 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.030655] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9e32602-72e5-42c9-a5ca-d66b2db4a656 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.043296] env[61987]: DEBUG nova.compute.provider_tree [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.063756] env[61987]: DEBUG nova.network.neutron [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.193173] env[61987]: DEBUG nova.network.neutron [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.409090] env[61987]: DEBUG nova.compute.manager [req-81463d75-f4cd-4cf8-a511-c90792f7d601 req-423a6dc7-a6f5-48b3-8eff-b3cd7113eb39 service nova] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Received event network-vif-deleted-1c4aa77c-6462-4435-bda8-10dc64c8e052 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 628.525664] env[61987]: DEBUG nova.compute.manager [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 628.548544] env[61987]: DEBUG nova.scheduler.client.report [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 628.563866] env[61987]: DEBUG nova.virt.hardware [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 628.564155] env[61987]: DEBUG nova.virt.hardware [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 628.564324] env[61987]: DEBUG nova.virt.hardware [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 628.564724] env[61987]: DEBUG nova.virt.hardware [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 628.564910] env[61987]: DEBUG nova.virt.hardware [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 628.565230] env[61987]: DEBUG nova.virt.hardware [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 628.565869] env[61987]: DEBUG nova.virt.hardware [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 628.566114] env[61987]: DEBUG nova.virt.hardware [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 628.566523] env[61987]: DEBUG nova.virt.hardware [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 628.566649] env[61987]: DEBUG nova.virt.hardware [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 628.567026] env[61987]: DEBUG nova.virt.hardware [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 628.568371] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c316cb98-9438-46aa-97cf-eb672aa80c80 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.579457] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e967e1-0420-41d9-9ce9-e82b90f6420d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.676640] env[61987]: ERROR nova.compute.manager [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port de0f490b-f7ab-4dfe-83b1-3574487c4556, please check neutron logs for more information. [ 628.676640] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 628.676640] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 628.676640] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 628.676640] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 628.676640] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 628.676640] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 628.676640] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 628.676640] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.676640] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 628.676640] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.676640] env[61987]: ERROR nova.compute.manager raise self.value [ 628.676640] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 628.676640] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 628.676640] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.676640] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 628.677063] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.677063] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 628.677063] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port de0f490b-f7ab-4dfe-83b1-3574487c4556, please check neutron logs for more information. [ 628.677063] env[61987]: ERROR nova.compute.manager [ 628.677063] env[61987]: Traceback (most recent call last): [ 628.677063] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 628.677063] env[61987]: listener.cb(fileno) [ 628.677063] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.677063] env[61987]: result = function(*args, **kwargs) [ 628.677063] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 628.677063] env[61987]: return func(*args, **kwargs) [ 628.677063] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 628.677063] env[61987]: raise e [ 628.677063] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 628.677063] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 628.677063] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 628.677063] env[61987]: created_port_ids = self._update_ports_for_instance( [ 628.677063] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 628.677063] env[61987]: with excutils.save_and_reraise_exception(): [ 628.677063] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.677063] env[61987]: self.force_reraise() [ 628.677063] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.677063] env[61987]: raise self.value [ 628.677063] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 628.677063] env[61987]: updated_port = self._update_port( [ 628.677063] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.677063] env[61987]: _ensure_no_port_binding_failure(port) [ 628.677063] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.677063] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 628.677756] env[61987]: nova.exception.PortBindingFailed: Binding failed for port de0f490b-f7ab-4dfe-83b1-3574487c4556, please check neutron logs for more information. [ 628.677756] env[61987]: Removing descriptor: 19 [ 628.677756] env[61987]: ERROR nova.compute.manager [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port de0f490b-f7ab-4dfe-83b1-3574487c4556, please check neutron logs for more information. [ 628.677756] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Traceback (most recent call last): [ 628.677756] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 628.677756] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] yield resources [ 628.677756] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 628.677756] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] self.driver.spawn(context, instance, image_meta, [ 628.677756] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 628.677756] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 628.677756] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 628.677756] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] vm_ref = self.build_virtual_machine(instance, [ 628.678039] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 628.678039] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] vif_infos = vmwarevif.get_vif_info(self._session, [ 628.678039] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 628.678039] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] for vif in network_info: [ 628.678039] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 628.678039] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] return self._sync_wrapper(fn, *args, **kwargs) [ 628.678039] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 628.678039] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] self.wait() [ 628.678039] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 628.678039] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] self[:] = self._gt.wait() [ 628.678039] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 628.678039] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] return self._exit_event.wait() [ 628.678039] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 628.678361] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] result = hub.switch() [ 628.678361] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 628.678361] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] return self.greenlet.switch() [ 628.678361] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.678361] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] result = function(*args, **kwargs) [ 628.678361] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 628.678361] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] return func(*args, **kwargs) [ 628.678361] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 628.678361] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] raise e [ 628.678361] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 628.678361] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] nwinfo = self.network_api.allocate_for_instance( [ 628.678361] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 628.678361] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] created_port_ids = self._update_ports_for_instance( [ 628.678648] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 628.678648] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] with excutils.save_and_reraise_exception(): [ 628.678648] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.678648] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] self.force_reraise() [ 628.678648] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.678648] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] raise self.value [ 628.678648] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 628.678648] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] updated_port = self._update_port( [ 628.678648] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.678648] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] _ensure_no_port_binding_failure(port) [ 628.678648] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.678648] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] raise exception.PortBindingFailed(port_id=port['id']) [ 628.678932] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] nova.exception.PortBindingFailed: Binding failed for port de0f490b-f7ab-4dfe-83b1-3574487c4556, please check neutron logs for more information. [ 628.678932] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] [ 628.678932] env[61987]: INFO nova.compute.manager [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Terminating instance [ 628.697565] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Releasing lock "refresh_cache-c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.698035] env[61987]: DEBUG nova.compute.manager [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 628.698237] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 628.698787] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fd3ea978-85b2-4a4b-87de-a5a7e142e9ad {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.708145] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-475fcce2-3ad1-45d4-b6bb-bdf380871e6c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.735538] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec could not be found. [ 628.735832] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 628.736041] env[61987]: INFO nova.compute.manager [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Took 0.04 seconds to destroy the instance on the hypervisor. [ 628.737270] env[61987]: DEBUG oslo.service.loopingcall [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 628.737270] env[61987]: DEBUG nova.compute.manager [-] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 628.737270] env[61987]: DEBUG nova.network.neutron [-] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 628.764290] env[61987]: DEBUG nova.network.neutron [-] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 629.056029] env[61987]: DEBUG oslo_concurrency.lockutils [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.577s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.057927] env[61987]: DEBUG nova.compute.manager [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 629.063499] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.542s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.065506] env[61987]: INFO nova.compute.claims [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 629.185997] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Acquiring lock "refresh_cache-967bc997-543f-4b35-968b-c530b82d9db0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.185997] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Acquired lock "refresh_cache-967bc997-543f-4b35-968b-c530b82d9db0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.185997] env[61987]: DEBUG nova.network.neutron [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 629.266310] env[61987]: DEBUG nova.network.neutron [-] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.570243] env[61987]: DEBUG nova.compute.utils [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 629.574770] env[61987]: DEBUG nova.compute.manager [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 629.575812] env[61987]: DEBUG nova.network.neutron [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 629.646560] env[61987]: DEBUG nova.policy [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b751f0742f054817bd824e94275b771c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8ddb948713643959b2baffa6d0de434', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 629.718599] env[61987]: DEBUG nova.network.neutron [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 629.771721] env[61987]: INFO nova.compute.manager [-] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Took 1.03 seconds to deallocate network for instance. [ 629.774832] env[61987]: DEBUG nova.compute.claims [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 629.775094] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.820633] env[61987]: DEBUG nova.network.neutron [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.064589] env[61987]: DEBUG nova.network.neutron [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Successfully created port: 3041ccf6-e87e-4e4e-8332-1a04de5361ce {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 630.087643] env[61987]: DEBUG nova.compute.manager [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 630.327845] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Releasing lock "refresh_cache-967bc997-543f-4b35-968b-c530b82d9db0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.327845] env[61987]: DEBUG nova.compute.manager [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 630.327845] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 630.327845] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a8c0a501-f941-49a1-a9d6-1bd8162d6520 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.335339] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f1aebb1-26cc-49a2-ac51-1ee2b0bc9a22 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.360462] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 967bc997-543f-4b35-968b-c530b82d9db0 could not be found. [ 630.360462] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 630.360605] env[61987]: INFO nova.compute.manager [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 630.360828] env[61987]: DEBUG oslo.service.loopingcall [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 630.363591] env[61987]: DEBUG nova.compute.manager [-] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 630.363808] env[61987]: DEBUG nova.network.neutron [-] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 630.390999] env[61987]: DEBUG nova.network.neutron [-] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.493492] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45775496-a52d-4589-afe0-074c4bdf4d9a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.501213] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c35be316-64ee-4cfc-871f-dc465d93dad7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.532399] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce6d2b4c-0b92-477f-a10b-1b3971737f3e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.539865] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb7d91cb-663e-418e-8b15-14cee7286bf3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.553738] env[61987]: DEBUG nova.compute.provider_tree [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.653924] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquiring lock "d038391c-2b32-455c-85d5-68221e7bd065" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.653924] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Lock "d038391c-2b32-455c-85d5-68221e7bd065" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.700437] env[61987]: DEBUG nova.compute.manager [req-c3d3f7d7-406e-49f5-8522-3156e0a091ea req-edef3ca1-4f0d-4234-b9e9-6f642babd66f service nova] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Received event network-changed-de0f490b-f7ab-4dfe-83b1-3574487c4556 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 630.700437] env[61987]: DEBUG nova.compute.manager [req-c3d3f7d7-406e-49f5-8522-3156e0a091ea req-edef3ca1-4f0d-4234-b9e9-6f642babd66f service nova] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Refreshing instance network info cache due to event network-changed-de0f490b-f7ab-4dfe-83b1-3574487c4556. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 630.700437] env[61987]: DEBUG oslo_concurrency.lockutils [req-c3d3f7d7-406e-49f5-8522-3156e0a091ea req-edef3ca1-4f0d-4234-b9e9-6f642babd66f service nova] Acquiring lock "refresh_cache-967bc997-543f-4b35-968b-c530b82d9db0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.700437] env[61987]: DEBUG oslo_concurrency.lockutils [req-c3d3f7d7-406e-49f5-8522-3156e0a091ea req-edef3ca1-4f0d-4234-b9e9-6f642babd66f service nova] Acquired lock "refresh_cache-967bc997-543f-4b35-968b-c530b82d9db0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.700437] env[61987]: DEBUG nova.network.neutron [req-c3d3f7d7-406e-49f5-8522-3156e0a091ea req-edef3ca1-4f0d-4234-b9e9-6f642babd66f service nova] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Refreshing network info cache for port de0f490b-f7ab-4dfe-83b1-3574487c4556 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 630.894149] env[61987]: DEBUG nova.network.neutron [-] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.056745] env[61987]: DEBUG nova.scheduler.client.report [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 631.101060] env[61987]: DEBUG nova.compute.manager [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 631.135729] env[61987]: DEBUG nova.virt.hardware [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 631.135977] env[61987]: DEBUG nova.virt.hardware [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 631.138710] env[61987]: DEBUG nova.virt.hardware [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 631.138959] env[61987]: DEBUG nova.virt.hardware [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 631.139122] env[61987]: DEBUG nova.virt.hardware [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 631.139280] env[61987]: DEBUG nova.virt.hardware [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 631.139513] env[61987]: DEBUG nova.virt.hardware [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 631.139676] env[61987]: DEBUG nova.virt.hardware [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 631.139847] env[61987]: DEBUG nova.virt.hardware [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 631.140019] env[61987]: DEBUG nova.virt.hardware [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 631.140195] env[61987]: DEBUG nova.virt.hardware [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 631.141107] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d235128c-b31f-4194-a4f0-d237eae8e129 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.153695] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e84eb07-7b26-4f89-8a23-fff5e6b3680b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.216810] env[61987]: DEBUG nova.network.neutron [req-c3d3f7d7-406e-49f5-8522-3156e0a091ea req-edef3ca1-4f0d-4234-b9e9-6f642babd66f service nova] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.322549] env[61987]: ERROR nova.compute.manager [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3041ccf6-e87e-4e4e-8332-1a04de5361ce, please check neutron logs for more information. [ 631.322549] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 631.322549] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 631.322549] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 631.322549] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 631.322549] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 631.322549] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 631.322549] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 631.322549] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.322549] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 631.322549] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.322549] env[61987]: ERROR nova.compute.manager raise self.value [ 631.322549] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 631.322549] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 631.322549] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.322549] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 631.323032] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.323032] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 631.323032] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3041ccf6-e87e-4e4e-8332-1a04de5361ce, please check neutron logs for more information. [ 631.323032] env[61987]: ERROR nova.compute.manager [ 631.324974] env[61987]: Traceback (most recent call last): [ 631.324974] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 631.324974] env[61987]: listener.cb(fileno) [ 631.324974] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.324974] env[61987]: result = function(*args, **kwargs) [ 631.324974] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 631.324974] env[61987]: return func(*args, **kwargs) [ 631.324974] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 631.324974] env[61987]: raise e [ 631.324974] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 631.324974] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 631.324974] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 631.324974] env[61987]: created_port_ids = self._update_ports_for_instance( [ 631.324974] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 631.324974] env[61987]: with excutils.save_and_reraise_exception(): [ 631.324974] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.324974] env[61987]: self.force_reraise() [ 631.324974] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.324974] env[61987]: raise self.value [ 631.324974] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 631.324974] env[61987]: updated_port = self._update_port( [ 631.324974] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.324974] env[61987]: _ensure_no_port_binding_failure(port) [ 631.324974] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.324974] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 631.324974] env[61987]: nova.exception.PortBindingFailed: Binding failed for port 3041ccf6-e87e-4e4e-8332-1a04de5361ce, please check neutron logs for more information. [ 631.324974] env[61987]: Removing descriptor: 19 [ 631.327957] env[61987]: ERROR nova.compute.manager [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3041ccf6-e87e-4e4e-8332-1a04de5361ce, please check neutron logs for more information. [ 631.327957] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Traceback (most recent call last): [ 631.327957] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 631.327957] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] yield resources [ 631.327957] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 631.327957] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] self.driver.spawn(context, instance, image_meta, [ 631.327957] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 631.327957] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.327957] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.327957] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] vm_ref = self.build_virtual_machine(instance, [ 631.327957] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.331208] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.331208] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.331208] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] for vif in network_info: [ 631.331208] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 631.331208] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] return self._sync_wrapper(fn, *args, **kwargs) [ 631.331208] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 631.331208] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] self.wait() [ 631.331208] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 631.331208] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] self[:] = self._gt.wait() [ 631.331208] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.331208] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] return self._exit_event.wait() [ 631.331208] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 631.331208] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] result = hub.switch() [ 631.331696] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 631.331696] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] return self.greenlet.switch() [ 631.331696] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.331696] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] result = function(*args, **kwargs) [ 631.331696] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 631.331696] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] return func(*args, **kwargs) [ 631.331696] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 631.331696] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] raise e [ 631.331696] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 631.331696] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] nwinfo = self.network_api.allocate_for_instance( [ 631.331696] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 631.331696] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] created_port_ids = self._update_ports_for_instance( [ 631.331696] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 631.332082] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] with excutils.save_and_reraise_exception(): [ 631.332082] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.332082] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] self.force_reraise() [ 631.332082] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.332082] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] raise self.value [ 631.332082] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 631.332082] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] updated_port = self._update_port( [ 631.332082] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.332082] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] _ensure_no_port_binding_failure(port) [ 631.332082] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.332082] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] raise exception.PortBindingFailed(port_id=port['id']) [ 631.332082] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] nova.exception.PortBindingFailed: Binding failed for port 3041ccf6-e87e-4e4e-8332-1a04de5361ce, please check neutron logs for more information. [ 631.332082] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] [ 631.332543] env[61987]: INFO nova.compute.manager [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Terminating instance [ 631.332543] env[61987]: DEBUG nova.network.neutron [req-c3d3f7d7-406e-49f5-8522-3156e0a091ea req-edef3ca1-4f0d-4234-b9e9-6f642babd66f service nova] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.398053] env[61987]: INFO nova.compute.manager [-] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Took 1.03 seconds to deallocate network for instance. [ 631.399038] env[61987]: DEBUG nova.compute.claims [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 631.399370] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.561505] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.498s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.564017] env[61987]: DEBUG nova.compute.manager [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 631.565783] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.841s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.566179] env[61987]: DEBUG nova.objects.instance [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Lazy-loading 'resources' on Instance uuid 6bf0d1f1-d079-4956-bc9a-86ba66885f3a {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 631.835063] env[61987]: DEBUG oslo_concurrency.lockutils [req-c3d3f7d7-406e-49f5-8522-3156e0a091ea req-edef3ca1-4f0d-4234-b9e9-6f642babd66f service nova] Releasing lock "refresh_cache-967bc997-543f-4b35-968b-c530b82d9db0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.835351] env[61987]: DEBUG nova.compute.manager [req-c3d3f7d7-406e-49f5-8522-3156e0a091ea req-edef3ca1-4f0d-4234-b9e9-6f642babd66f service nova] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Received event network-vif-deleted-de0f490b-f7ab-4dfe-83b1-3574487c4556 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 631.839173] env[61987]: DEBUG oslo_concurrency.lockutils [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Acquiring lock "refresh_cache-ab2edc70-b4c6-4db3-b0d6-53148764efdd" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.839173] env[61987]: DEBUG oslo_concurrency.lockutils [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Acquired lock "refresh_cache-ab2edc70-b4c6-4db3-b0d6-53148764efdd" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.839173] env[61987]: DEBUG nova.network.neutron [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 632.070060] env[61987]: DEBUG nova.compute.utils [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 632.075782] env[61987]: DEBUG nova.compute.manager [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 632.076123] env[61987]: DEBUG nova.network.neutron [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 632.152602] env[61987]: DEBUG nova.policy [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a819a29f88444630b7adc2cac30e71e2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1acc92eb0c4048789197010b200d9e35', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 632.372298] env[61987]: DEBUG nova.network.neutron [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 632.527863] env[61987]: DEBUG nova.network.neutron [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.557701] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3787866c-c3b1-4727-a829-5c0d7689b088 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.566852] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144e89a3-e343-49c5-a043-d7d120e8f18d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.598623] env[61987]: DEBUG nova.compute.manager [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 632.601855] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b8d09f-6023-4f96-9c29-939ed336a5ca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.612023] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d245dacb-96f8-44dc-a020-ebcdfed3d846 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.624758] env[61987]: DEBUG nova.compute.provider_tree [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.645935] env[61987]: DEBUG nova.network.neutron [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Successfully created port: 264cbccf-a94c-47d7-a785-9b2b8bd0fa58 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 632.726806] env[61987]: DEBUG nova.compute.manager [req-ca055a31-5ae4-4d62-a084-dc67f382da40 req-f9da8ed8-03e2-4657-a596-a7495de936de service nova] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Received event network-changed-3041ccf6-e87e-4e4e-8332-1a04de5361ce {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 632.727015] env[61987]: DEBUG nova.compute.manager [req-ca055a31-5ae4-4d62-a084-dc67f382da40 req-f9da8ed8-03e2-4657-a596-a7495de936de service nova] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Refreshing instance network info cache due to event network-changed-3041ccf6-e87e-4e4e-8332-1a04de5361ce. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 632.727408] env[61987]: DEBUG oslo_concurrency.lockutils [req-ca055a31-5ae4-4d62-a084-dc67f382da40 req-f9da8ed8-03e2-4657-a596-a7495de936de service nova] Acquiring lock "refresh_cache-ab2edc70-b4c6-4db3-b0d6-53148764efdd" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.033508] env[61987]: DEBUG oslo_concurrency.lockutils [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Releasing lock "refresh_cache-ab2edc70-b4c6-4db3-b0d6-53148764efdd" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.033946] env[61987]: DEBUG nova.compute.manager [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 633.034159] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 633.034483] env[61987]: DEBUG oslo_concurrency.lockutils [req-ca055a31-5ae4-4d62-a084-dc67f382da40 req-f9da8ed8-03e2-4657-a596-a7495de936de service nova] Acquired lock "refresh_cache-ab2edc70-b4c6-4db3-b0d6-53148764efdd" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.034661] env[61987]: DEBUG nova.network.neutron [req-ca055a31-5ae4-4d62-a084-dc67f382da40 req-f9da8ed8-03e2-4657-a596-a7495de936de service nova] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Refreshing network info cache for port 3041ccf6-e87e-4e4e-8332-1a04de5361ce {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 633.035791] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2b673832-f8af-419b-8675-d31f2b1f9ab0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.045579] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da193b6e-e0fe-4b75-8cee-377d850c9ff8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.068691] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ab2edc70-b4c6-4db3-b0d6-53148764efdd could not be found. [ 633.068804] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 633.068986] env[61987]: INFO nova.compute.manager [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Took 0.03 seconds to destroy the instance on the hypervisor. [ 633.069269] env[61987]: DEBUG oslo.service.loopingcall [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 633.069495] env[61987]: DEBUG nova.compute.manager [-] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 633.069589] env[61987]: DEBUG nova.network.neutron [-] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 633.088061] env[61987]: DEBUG nova.network.neutron [-] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 633.127563] env[61987]: DEBUG nova.scheduler.client.report [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 633.597035] env[61987]: DEBUG nova.network.neutron [-] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.601647] env[61987]: DEBUG nova.network.neutron [req-ca055a31-5ae4-4d62-a084-dc67f382da40 req-f9da8ed8-03e2-4657-a596-a7495de936de service nova] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 633.610595] env[61987]: DEBUG nova.compute.manager [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 633.635106] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.069s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.637778] env[61987]: DEBUG oslo_concurrency.lockutils [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.870s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.640244] env[61987]: INFO nova.compute.claims [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 633.665733] env[61987]: DEBUG nova.virt.hardware [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 633.665733] env[61987]: DEBUG nova.virt.hardware [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 633.665733] env[61987]: DEBUG nova.virt.hardware [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 633.666008] env[61987]: DEBUG nova.virt.hardware [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 633.666008] env[61987]: DEBUG nova.virt.hardware [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 633.666181] env[61987]: DEBUG nova.virt.hardware [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 633.666352] env[61987]: DEBUG nova.virt.hardware [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 633.666514] env[61987]: DEBUG nova.virt.hardware [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 633.667606] env[61987]: DEBUG nova.virt.hardware [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 633.667658] env[61987]: DEBUG nova.virt.hardware [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 633.667830] env[61987]: DEBUG nova.virt.hardware [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 633.669021] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e56dbb28-16b5-4516-99e0-0e5c077d46a7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.673506] env[61987]: INFO nova.scheduler.client.report [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Deleted allocations for instance 6bf0d1f1-d079-4956-bc9a-86ba66885f3a [ 633.684521] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef4f9c2c-b066-4550-96b7-10e9f7a4da36 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.763373] env[61987]: DEBUG nova.network.neutron [req-ca055a31-5ae4-4d62-a084-dc67f382da40 req-f9da8ed8-03e2-4657-a596-a7495de936de service nova] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.952302] env[61987]: ERROR nova.compute.manager [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 264cbccf-a94c-47d7-a785-9b2b8bd0fa58, please check neutron logs for more information. [ 633.952302] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 633.952302] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 633.952302] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 633.952302] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.952302] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 633.952302] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.952302] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 633.952302] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.952302] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 633.952302] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.952302] env[61987]: ERROR nova.compute.manager raise self.value [ 633.952302] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.952302] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 633.952302] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.952302] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 633.953165] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.953165] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 633.953165] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 264cbccf-a94c-47d7-a785-9b2b8bd0fa58, please check neutron logs for more information. [ 633.953165] env[61987]: ERROR nova.compute.manager [ 633.953165] env[61987]: Traceback (most recent call last): [ 633.953165] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 633.953165] env[61987]: listener.cb(fileno) [ 633.953165] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.953165] env[61987]: result = function(*args, **kwargs) [ 633.953165] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 633.953165] env[61987]: return func(*args, **kwargs) [ 633.953165] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 633.953165] env[61987]: raise e [ 633.953165] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 633.953165] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 633.953165] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.953165] env[61987]: created_port_ids = self._update_ports_for_instance( [ 633.953165] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.953165] env[61987]: with excutils.save_and_reraise_exception(): [ 633.953165] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.953165] env[61987]: self.force_reraise() [ 633.953165] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.953165] env[61987]: raise self.value [ 633.953165] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.953165] env[61987]: updated_port = self._update_port( [ 633.953165] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.953165] env[61987]: _ensure_no_port_binding_failure(port) [ 633.953165] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.953165] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 633.953881] env[61987]: nova.exception.PortBindingFailed: Binding failed for port 264cbccf-a94c-47d7-a785-9b2b8bd0fa58, please check neutron logs for more information. [ 633.953881] env[61987]: Removing descriptor: 18 [ 633.953881] env[61987]: ERROR nova.compute.manager [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 264cbccf-a94c-47d7-a785-9b2b8bd0fa58, please check neutron logs for more information. [ 633.953881] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Traceback (most recent call last): [ 633.953881] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 633.953881] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] yield resources [ 633.953881] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 633.953881] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] self.driver.spawn(context, instance, image_meta, [ 633.953881] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 633.953881] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 633.953881] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 633.953881] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] vm_ref = self.build_virtual_machine(instance, [ 633.954210] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 633.954210] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] vif_infos = vmwarevif.get_vif_info(self._session, [ 633.954210] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 633.954210] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] for vif in network_info: [ 633.954210] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 633.954210] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] return self._sync_wrapper(fn, *args, **kwargs) [ 633.954210] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 633.954210] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] self.wait() [ 633.954210] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 633.954210] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] self[:] = self._gt.wait() [ 633.954210] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 633.954210] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] return self._exit_event.wait() [ 633.954210] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 633.954539] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] result = hub.switch() [ 633.954539] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 633.954539] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] return self.greenlet.switch() [ 633.954539] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.954539] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] result = function(*args, **kwargs) [ 633.954539] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 633.954539] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] return func(*args, **kwargs) [ 633.954539] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 633.954539] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] raise e [ 633.954539] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 633.954539] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] nwinfo = self.network_api.allocate_for_instance( [ 633.954539] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 633.954539] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] created_port_ids = self._update_ports_for_instance( [ 633.954855] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 633.954855] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] with excutils.save_and_reraise_exception(): [ 633.954855] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.954855] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] self.force_reraise() [ 633.954855] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.954855] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] raise self.value [ 633.954855] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 633.954855] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] updated_port = self._update_port( [ 633.954855] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.954855] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] _ensure_no_port_binding_failure(port) [ 633.954855] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.954855] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] raise exception.PortBindingFailed(port_id=port['id']) [ 633.958134] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] nova.exception.PortBindingFailed: Binding failed for port 264cbccf-a94c-47d7-a785-9b2b8bd0fa58, please check neutron logs for more information. [ 633.958134] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] [ 633.958134] env[61987]: INFO nova.compute.manager [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Terminating instance [ 634.075089] env[61987]: DEBUG oslo_concurrency.lockutils [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "ad40937f-d5a9-4708-8b81-06087d38d765" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.075089] env[61987]: DEBUG oslo_concurrency.lockutils [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ad40937f-d5a9-4708-8b81-06087d38d765" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.100314] env[61987]: INFO nova.compute.manager [-] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Took 1.03 seconds to deallocate network for instance. [ 634.103255] env[61987]: DEBUG nova.compute.claims [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 634.103389] env[61987]: DEBUG oslo_concurrency.lockutils [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.184015] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f820e7c0-004d-45c2-9c5c-75ec8c8ce9db tempest-ServerDiagnosticsV248Test-33428269 tempest-ServerDiagnosticsV248Test-33428269-project-member] Lock "6bf0d1f1-d079-4956-bc9a-86ba66885f3a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.763s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.266337] env[61987]: DEBUG oslo_concurrency.lockutils [req-ca055a31-5ae4-4d62-a084-dc67f382da40 req-f9da8ed8-03e2-4657-a596-a7495de936de service nova] Releasing lock "refresh_cache-ab2edc70-b4c6-4db3-b0d6-53148764efdd" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.266759] env[61987]: DEBUG nova.compute.manager [req-ca055a31-5ae4-4d62-a084-dc67f382da40 req-f9da8ed8-03e2-4657-a596-a7495de936de service nova] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Received event network-vif-deleted-3041ccf6-e87e-4e4e-8332-1a04de5361ce {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 634.457387] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Acquiring lock "refresh_cache-388b62aa-6647-4aca-acb2-8d7d7d424c8c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.457499] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Acquired lock "refresh_cache-388b62aa-6647-4aca-acb2-8d7d7d424c8c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.458084] env[61987]: DEBUG nova.network.neutron [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 634.754831] env[61987]: DEBUG nova.compute.manager [req-61676f8c-67da-44e0-9b67-8b0e7a892a7e req-22e0ba59-4ba4-4211-88e2-696eaccf1c63 service nova] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Received event network-changed-264cbccf-a94c-47d7-a785-9b2b8bd0fa58 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 634.754831] env[61987]: DEBUG nova.compute.manager [req-61676f8c-67da-44e0-9b67-8b0e7a892a7e req-22e0ba59-4ba4-4211-88e2-696eaccf1c63 service nova] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Refreshing instance network info cache due to event network-changed-264cbccf-a94c-47d7-a785-9b2b8bd0fa58. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 634.754831] env[61987]: DEBUG oslo_concurrency.lockutils [req-61676f8c-67da-44e0-9b67-8b0e7a892a7e req-22e0ba59-4ba4-4211-88e2-696eaccf1c63 service nova] Acquiring lock "refresh_cache-388b62aa-6647-4aca-acb2-8d7d7d424c8c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.981108] env[61987]: DEBUG nova.network.neutron [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 635.078343] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a8a8c3d-d10a-4f23-82a8-3508b48962f5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.086100] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e9d355-7ad3-4197-9491-2182e3f47e69 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.117247] env[61987]: DEBUG nova.network.neutron [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.118900] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19f518c-adca-4ac1-9e0a-9dfcb5d83f38 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.128928] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf4b2cf-d529-4226-9004-38312ec1c587 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.143276] env[61987]: DEBUG nova.compute.provider_tree [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.624803] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Releasing lock "refresh_cache-388b62aa-6647-4aca-acb2-8d7d7d424c8c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.625355] env[61987]: DEBUG nova.compute.manager [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 635.625587] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 635.625950] env[61987]: DEBUG oslo_concurrency.lockutils [req-61676f8c-67da-44e0-9b67-8b0e7a892a7e req-22e0ba59-4ba4-4211-88e2-696eaccf1c63 service nova] Acquired lock "refresh_cache-388b62aa-6647-4aca-acb2-8d7d7d424c8c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.626196] env[61987]: DEBUG nova.network.neutron [req-61676f8c-67da-44e0-9b67-8b0e7a892a7e req-22e0ba59-4ba4-4211-88e2-696eaccf1c63 service nova] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Refreshing network info cache for port 264cbccf-a94c-47d7-a785-9b2b8bd0fa58 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 635.627429] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-534846b7-19d5-4ed4-9f59-d58d2b67149d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.637771] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b469ca-7b57-4bc9-b712-0d9801ccb795 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.649182] env[61987]: DEBUG nova.scheduler.client.report [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 635.663805] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 388b62aa-6647-4aca-acb2-8d7d7d424c8c could not be found. [ 635.664059] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 635.664357] env[61987]: INFO nova.compute.manager [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 635.664613] env[61987]: DEBUG oslo.service.loopingcall [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 635.664847] env[61987]: DEBUG nova.compute.manager [-] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 635.664938] env[61987]: DEBUG nova.network.neutron [-] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 635.680494] env[61987]: DEBUG nova.network.neutron [-] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.146811] env[61987]: DEBUG nova.network.neutron [req-61676f8c-67da-44e0-9b67-8b0e7a892a7e req-22e0ba59-4ba4-4211-88e2-696eaccf1c63 service nova] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.148804] env[61987]: DEBUG oslo_concurrency.lockutils [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Acquiring lock "1afcffda-0bd2-4cbe-8cfb-12a91bb50975" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.149036] env[61987]: DEBUG oslo_concurrency.lockutils [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Lock "1afcffda-0bd2-4cbe-8cfb-12a91bb50975" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.154068] env[61987]: DEBUG oslo_concurrency.lockutils [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.517s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.154535] env[61987]: DEBUG nova.compute.manager [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 636.157158] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.180s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.158366] env[61987]: INFO nova.compute.claims [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 636.182969] env[61987]: DEBUG nova.network.neutron [-] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.252158] env[61987]: DEBUG nova.network.neutron [req-61676f8c-67da-44e0-9b67-8b0e7a892a7e req-22e0ba59-4ba4-4211-88e2-696eaccf1c63 service nova] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.663662] env[61987]: DEBUG nova.compute.utils [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 636.666785] env[61987]: DEBUG nova.compute.manager [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 636.666953] env[61987]: DEBUG nova.network.neutron [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 636.686142] env[61987]: INFO nova.compute.manager [-] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Took 1.02 seconds to deallocate network for instance. [ 636.689217] env[61987]: DEBUG nova.compute.claims [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 636.689392] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.706179] env[61987]: DEBUG nova.policy [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dbd4a31a68334b179d95c3a864b4dd99', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9d580fa0eb834f8f90860727e62a252d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 636.754676] env[61987]: DEBUG oslo_concurrency.lockutils [req-61676f8c-67da-44e0-9b67-8b0e7a892a7e req-22e0ba59-4ba4-4211-88e2-696eaccf1c63 service nova] Releasing lock "refresh_cache-388b62aa-6647-4aca-acb2-8d7d7d424c8c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.754939] env[61987]: DEBUG nova.compute.manager [req-61676f8c-67da-44e0-9b67-8b0e7a892a7e req-22e0ba59-4ba4-4211-88e2-696eaccf1c63 service nova] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Received event network-vif-deleted-264cbccf-a94c-47d7-a785-9b2b8bd0fa58 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 637.075650] env[61987]: DEBUG nova.network.neutron [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Successfully created port: 41fd9385-de81-4bd6-8feb-55408aca0c0a {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 637.167559] env[61987]: DEBUG nova.compute.manager [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 637.563321] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81214618-a402-4854-90f5-087680531260 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.570939] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86a629c4-4a0d-4a40-8c34-9f88ce44a3e8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.600311] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d86e2387-b205-46af-a923-7fb0a399ef5e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.607510] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de3a6763-b063-4a90-8c09-20b5dd0c339e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.623100] env[61987]: DEBUG nova.compute.provider_tree [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.880758] env[61987]: DEBUG nova.compute.manager [req-d10cab09-55b7-454f-8148-34901c26b0f8 req-88fc43aa-474a-4f30-8e76-08b93c9986a0 service nova] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Received event network-changed-41fd9385-de81-4bd6-8feb-55408aca0c0a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 637.880965] env[61987]: DEBUG nova.compute.manager [req-d10cab09-55b7-454f-8148-34901c26b0f8 req-88fc43aa-474a-4f30-8e76-08b93c9986a0 service nova] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Refreshing instance network info cache due to event network-changed-41fd9385-de81-4bd6-8feb-55408aca0c0a. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 637.881237] env[61987]: DEBUG oslo_concurrency.lockutils [req-d10cab09-55b7-454f-8148-34901c26b0f8 req-88fc43aa-474a-4f30-8e76-08b93c9986a0 service nova] Acquiring lock "refresh_cache-96bbc732-cb58-4de2-80d3-c527228adaea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.881421] env[61987]: DEBUG oslo_concurrency.lockutils [req-d10cab09-55b7-454f-8148-34901c26b0f8 req-88fc43aa-474a-4f30-8e76-08b93c9986a0 service nova] Acquired lock "refresh_cache-96bbc732-cb58-4de2-80d3-c527228adaea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.881634] env[61987]: DEBUG nova.network.neutron [req-d10cab09-55b7-454f-8148-34901c26b0f8 req-88fc43aa-474a-4f30-8e76-08b93c9986a0 service nova] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Refreshing network info cache for port 41fd9385-de81-4bd6-8feb-55408aca0c0a {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 638.127367] env[61987]: DEBUG nova.scheduler.client.report [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 638.179924] env[61987]: DEBUG nova.compute.manager [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 638.182934] env[61987]: ERROR nova.compute.manager [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 41fd9385-de81-4bd6-8feb-55408aca0c0a, please check neutron logs for more information. [ 638.182934] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 638.182934] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 638.182934] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 638.182934] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 638.182934] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 638.182934] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 638.182934] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 638.182934] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.182934] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 638.182934] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.182934] env[61987]: ERROR nova.compute.manager raise self.value [ 638.182934] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 638.182934] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 638.182934] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.182934] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 638.183518] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.183518] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 638.183518] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 41fd9385-de81-4bd6-8feb-55408aca0c0a, please check neutron logs for more information. [ 638.183518] env[61987]: ERROR nova.compute.manager [ 638.183518] env[61987]: Traceback (most recent call last): [ 638.183518] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 638.183518] env[61987]: listener.cb(fileno) [ 638.183518] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.183518] env[61987]: result = function(*args, **kwargs) [ 638.183518] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 638.183518] env[61987]: return func(*args, **kwargs) [ 638.183518] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 638.183518] env[61987]: raise e [ 638.183518] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 638.183518] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 638.183518] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 638.183518] env[61987]: created_port_ids = self._update_ports_for_instance( [ 638.183518] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 638.183518] env[61987]: with excutils.save_and_reraise_exception(): [ 638.183518] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.183518] env[61987]: self.force_reraise() [ 638.183518] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.183518] env[61987]: raise self.value [ 638.183518] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 638.183518] env[61987]: updated_port = self._update_port( [ 638.183518] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.183518] env[61987]: _ensure_no_port_binding_failure(port) [ 638.183518] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.183518] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 638.184300] env[61987]: nova.exception.PortBindingFailed: Binding failed for port 41fd9385-de81-4bd6-8feb-55408aca0c0a, please check neutron logs for more information. [ 638.184300] env[61987]: Removing descriptor: 18 [ 638.207667] env[61987]: DEBUG nova.virt.hardware [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 638.207920] env[61987]: DEBUG nova.virt.hardware [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 638.208089] env[61987]: DEBUG nova.virt.hardware [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 638.208276] env[61987]: DEBUG nova.virt.hardware [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 638.208421] env[61987]: DEBUG nova.virt.hardware [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 638.208566] env[61987]: DEBUG nova.virt.hardware [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 638.208764] env[61987]: DEBUG nova.virt.hardware [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 638.208919] env[61987]: DEBUG nova.virt.hardware [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 638.209118] env[61987]: DEBUG nova.virt.hardware [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 638.209289] env[61987]: DEBUG nova.virt.hardware [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 638.209460] env[61987]: DEBUG nova.virt.hardware [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 638.210354] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70dd14aa-13cc-4cf2-878e-e3ad5ee5ffc1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.220233] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f3294d8-6049-4460-9ff0-a9e2cfc24867 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.233603] env[61987]: ERROR nova.compute.manager [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 41fd9385-de81-4bd6-8feb-55408aca0c0a, please check neutron logs for more information. [ 638.233603] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Traceback (most recent call last): [ 638.233603] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 638.233603] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] yield resources [ 638.233603] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 638.233603] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] self.driver.spawn(context, instance, image_meta, [ 638.233603] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 638.233603] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.233603] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.233603] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] vm_ref = self.build_virtual_machine(instance, [ 638.233603] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.233941] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.233941] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.233941] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] for vif in network_info: [ 638.233941] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 638.233941] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] return self._sync_wrapper(fn, *args, **kwargs) [ 638.233941] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 638.233941] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] self.wait() [ 638.233941] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 638.233941] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] self[:] = self._gt.wait() [ 638.233941] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.233941] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] return self._exit_event.wait() [ 638.233941] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 638.233941] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] current.throw(*self._exc) [ 638.234320] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.234320] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] result = function(*args, **kwargs) [ 638.234320] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 638.234320] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] return func(*args, **kwargs) [ 638.234320] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 638.234320] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] raise e [ 638.234320] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 638.234320] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] nwinfo = self.network_api.allocate_for_instance( [ 638.234320] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 638.234320] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] created_port_ids = self._update_ports_for_instance( [ 638.234320] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 638.234320] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] with excutils.save_and_reraise_exception(): [ 638.234320] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.234685] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] self.force_reraise() [ 638.234685] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.234685] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] raise self.value [ 638.234685] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 638.234685] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] updated_port = self._update_port( [ 638.234685] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.234685] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] _ensure_no_port_binding_failure(port) [ 638.234685] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.234685] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] raise exception.PortBindingFailed(port_id=port['id']) [ 638.234685] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] nova.exception.PortBindingFailed: Binding failed for port 41fd9385-de81-4bd6-8feb-55408aca0c0a, please check neutron logs for more information. [ 638.234685] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] [ 638.234685] env[61987]: INFO nova.compute.manager [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Terminating instance [ 638.399349] env[61987]: DEBUG nova.network.neutron [req-d10cab09-55b7-454f-8148-34901c26b0f8 req-88fc43aa-474a-4f30-8e76-08b93c9986a0 service nova] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 638.519810] env[61987]: DEBUG nova.network.neutron [req-d10cab09-55b7-454f-8148-34901c26b0f8 req-88fc43aa-474a-4f30-8e76-08b93c9986a0 service nova] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.632965] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.476s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.633567] env[61987]: DEBUG nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 638.636283] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.159s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.638056] env[61987]: INFO nova.compute.claims [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 638.738695] env[61987]: DEBUG oslo_concurrency.lockutils [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Acquiring lock "refresh_cache-96bbc732-cb58-4de2-80d3-c527228adaea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.022997] env[61987]: DEBUG oslo_concurrency.lockutils [req-d10cab09-55b7-454f-8148-34901c26b0f8 req-88fc43aa-474a-4f30-8e76-08b93c9986a0 service nova] Releasing lock "refresh_cache-96bbc732-cb58-4de2-80d3-c527228adaea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.023395] env[61987]: DEBUG oslo_concurrency.lockutils [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Acquired lock "refresh_cache-96bbc732-cb58-4de2-80d3-c527228adaea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.023581] env[61987]: DEBUG nova.network.neutron [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 639.142568] env[61987]: DEBUG nova.compute.utils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 639.146296] env[61987]: DEBUG nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 639.146487] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 639.198388] env[61987]: DEBUG nova.policy [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f9b90037d9384f549640bfcf344f0669', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '69749b752d1d4ad087f70f3d2ebe7b10', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 639.547804] env[61987]: DEBUG nova.network.neutron [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 639.647788] env[61987]: DEBUG nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 639.684404] env[61987]: DEBUG nova.network.neutron [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.746444] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Successfully created port: 59570761-d84d-4350-9f6c-17502c6e58ff {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 639.907675] env[61987]: DEBUG nova.compute.manager [req-b8e38cba-44ca-4a0e-8776-9cd1bcb3a1b7 req-df8251ae-c9a7-4671-b48b-dc7459e01e2c service nova] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Received event network-vif-deleted-41fd9385-de81-4bd6-8feb-55408aca0c0a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 640.060104] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2eab4ef-50fa-47b5-b300-aebe27d59715 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.068814] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-165ca2d0-8d8f-487a-8a8d-44e54360395d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.100205] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7fd7109-0fe1-4525-bdb7-8c903ea57764 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.108120] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d55d0c5a-7884-4694-a9ed-7e54586b9aaf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.125154] env[61987]: DEBUG nova.compute.provider_tree [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 640.189588] env[61987]: DEBUG oslo_concurrency.lockutils [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Releasing lock "refresh_cache-96bbc732-cb58-4de2-80d3-c527228adaea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.189588] env[61987]: DEBUG nova.compute.manager [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 640.189588] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 640.189588] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-99394ce0-53d3-47b9-86ae-7a1a282efddb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.201372] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82ef5ab0-872a-49e4-981f-e10bed0ab4e2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.224125] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 96bbc732-cb58-4de2-80d3-c527228adaea could not be found. [ 640.224357] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 640.224536] env[61987]: INFO nova.compute.manager [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Took 0.04 seconds to destroy the instance on the hypervisor. [ 640.224773] env[61987]: DEBUG oslo.service.loopingcall [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 640.224987] env[61987]: DEBUG nova.compute.manager [-] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 640.225090] env[61987]: DEBUG nova.network.neutron [-] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 640.248019] env[61987]: DEBUG nova.network.neutron [-] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 640.632802] env[61987]: DEBUG nova.scheduler.client.report [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 640.661219] env[61987]: DEBUG nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 640.693319] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 640.693319] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 640.693319] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 640.693792] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 640.694073] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 640.694349] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 640.694660] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 640.694919] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 640.695214] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 640.696022] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 640.696022] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 640.696642] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70168790-0996-417c-b4c8-1f469ad0203e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.705848] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e28c1307-5cd4-4915-8919-5ad6c2ef1f15 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.749694] env[61987]: DEBUG nova.network.neutron [-] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.935248] env[61987]: ERROR nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 59570761-d84d-4350-9f6c-17502c6e58ff, please check neutron logs for more information. [ 640.935248] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 640.935248] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 640.935248] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 640.935248] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 640.935248] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 640.935248] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 640.935248] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 640.935248] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.935248] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 640.935248] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.935248] env[61987]: ERROR nova.compute.manager raise self.value [ 640.935248] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 640.935248] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 640.935248] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.935248] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 640.935673] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.935673] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 640.935673] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 59570761-d84d-4350-9f6c-17502c6e58ff, please check neutron logs for more information. [ 640.935673] env[61987]: ERROR nova.compute.manager [ 640.935673] env[61987]: Traceback (most recent call last): [ 640.935673] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 640.935673] env[61987]: listener.cb(fileno) [ 640.935673] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.935673] env[61987]: result = function(*args, **kwargs) [ 640.935673] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.935673] env[61987]: return func(*args, **kwargs) [ 640.935673] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 640.935673] env[61987]: raise e [ 640.935673] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 640.935673] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 640.935673] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 640.935673] env[61987]: created_port_ids = self._update_ports_for_instance( [ 640.935673] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 640.935673] env[61987]: with excutils.save_and_reraise_exception(): [ 640.935673] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.935673] env[61987]: self.force_reraise() [ 640.935673] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.935673] env[61987]: raise self.value [ 640.935673] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 640.935673] env[61987]: updated_port = self._update_port( [ 640.935673] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.935673] env[61987]: _ensure_no_port_binding_failure(port) [ 640.935673] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.935673] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 640.936383] env[61987]: nova.exception.PortBindingFailed: Binding failed for port 59570761-d84d-4350-9f6c-17502c6e58ff, please check neutron logs for more information. [ 640.936383] env[61987]: Removing descriptor: 18 [ 640.936383] env[61987]: ERROR nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 59570761-d84d-4350-9f6c-17502c6e58ff, please check neutron logs for more information. [ 640.936383] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Traceback (most recent call last): [ 640.936383] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 640.936383] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] yield resources [ 640.936383] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 640.936383] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] self.driver.spawn(context, instance, image_meta, [ 640.936383] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 640.936383] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] self._vmops.spawn(context, instance, image_meta, injected_files, [ 640.936383] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 640.936383] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] vm_ref = self.build_virtual_machine(instance, [ 640.936744] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 640.936744] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] vif_infos = vmwarevif.get_vif_info(self._session, [ 640.936744] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 640.936744] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] for vif in network_info: [ 640.936744] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 640.936744] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] return self._sync_wrapper(fn, *args, **kwargs) [ 640.936744] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 640.936744] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] self.wait() [ 640.936744] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 640.936744] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] self[:] = self._gt.wait() [ 640.936744] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 640.936744] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] return self._exit_event.wait() [ 640.936744] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 640.937085] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] result = hub.switch() [ 640.937085] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 640.937085] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] return self.greenlet.switch() [ 640.937085] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.937085] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] result = function(*args, **kwargs) [ 640.937085] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.937085] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] return func(*args, **kwargs) [ 640.937085] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 640.937085] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] raise e [ 640.937085] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 640.937085] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] nwinfo = self.network_api.allocate_for_instance( [ 640.937085] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 640.937085] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] created_port_ids = self._update_ports_for_instance( [ 640.937433] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 640.937433] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] with excutils.save_and_reraise_exception(): [ 640.937433] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.937433] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] self.force_reraise() [ 640.937433] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.937433] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] raise self.value [ 640.937433] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 640.937433] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] updated_port = self._update_port( [ 640.937433] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.937433] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] _ensure_no_port_binding_failure(port) [ 640.937433] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.937433] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] raise exception.PortBindingFailed(port_id=port['id']) [ 640.937701] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] nova.exception.PortBindingFailed: Binding failed for port 59570761-d84d-4350-9f6c-17502c6e58ff, please check neutron logs for more information. [ 640.937701] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] [ 640.937701] env[61987]: INFO nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Terminating instance [ 641.139688] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.503s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.139688] env[61987]: DEBUG nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 641.142101] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.199s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.143552] env[61987]: INFO nova.compute.claims [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 641.252557] env[61987]: INFO nova.compute.manager [-] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Took 1.03 seconds to deallocate network for instance. [ 641.255308] env[61987]: DEBUG nova.compute.claims [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 641.255491] env[61987]: DEBUG oslo_concurrency.lockutils [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.361996] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Acquiring lock "95641d0b-970c-4b94-8568-5c46d0808345" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.362251] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Lock "95641d0b-970c-4b94-8568-5c46d0808345" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.440825] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "refresh_cache-699c853a-a88f-4890-b712-00f00a6c9838" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 641.440950] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquired lock "refresh_cache-699c853a-a88f-4890-b712-00f00a6c9838" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.441147] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 641.648606] env[61987]: DEBUG nova.compute.utils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 641.652256] env[61987]: DEBUG nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 641.652433] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 641.697859] env[61987]: DEBUG nova.policy [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f9b90037d9384f549640bfcf344f0669', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '69749b752d1d4ad087f70f3d2ebe7b10', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 641.972136] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 641.975611] env[61987]: DEBUG nova.compute.manager [req-c50229fa-5ca3-4ae8-b3c0-13ecaa17963a req-a736b411-0ad1-4741-a02b-90fa5800c7a0 service nova] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Received event network-changed-59570761-d84d-4350-9f6c-17502c6e58ff {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 641.976489] env[61987]: DEBUG nova.compute.manager [req-c50229fa-5ca3-4ae8-b3c0-13ecaa17963a req-a736b411-0ad1-4741-a02b-90fa5800c7a0 service nova] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Refreshing instance network info cache due to event network-changed-59570761-d84d-4350-9f6c-17502c6e58ff. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 641.976800] env[61987]: DEBUG oslo_concurrency.lockutils [req-c50229fa-5ca3-4ae8-b3c0-13ecaa17963a req-a736b411-0ad1-4741-a02b-90fa5800c7a0 service nova] Acquiring lock "refresh_cache-699c853a-a88f-4890-b712-00f00a6c9838" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.042286] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Successfully created port: 26ebdb7d-365e-4aa3-a0ae-35812dd80bbb {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 642.120638] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.153514] env[61987]: DEBUG nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 642.576926] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c57aa5-7660-4ee9-9b15-be083adc27c5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.586307] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc7d575c-12d8-402b-944e-0f3cb3ea8f93 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.616784] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af2a9c3-f182-42a5-914b-a1c748b1dd17 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.623818] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Releasing lock "refresh_cache-699c853a-a88f-4890-b712-00f00a6c9838" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.624305] env[61987]: DEBUG nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 642.624589] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 642.624961] env[61987]: DEBUG oslo_concurrency.lockutils [req-c50229fa-5ca3-4ae8-b3c0-13ecaa17963a req-a736b411-0ad1-4741-a02b-90fa5800c7a0 service nova] Acquired lock "refresh_cache-699c853a-a88f-4890-b712-00f00a6c9838" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.625254] env[61987]: DEBUG nova.network.neutron [req-c50229fa-5ca3-4ae8-b3c0-13ecaa17963a req-a736b411-0ad1-4741-a02b-90fa5800c7a0 service nova] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Refreshing network info cache for port 59570761-d84d-4350-9f6c-17502c6e58ff {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 642.626367] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-83df62f4-d0a9-4d1e-84a0-bd455a4ab1f2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.629542] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d2fb70-3489-40e6-a4da-4dffbcb5e848 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.648392] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d99ee7-9237-448d-94db-8f6db9955bb8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.659519] env[61987]: DEBUG nova.compute.provider_tree [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 642.676384] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 699c853a-a88f-4890-b712-00f00a6c9838 could not be found. [ 642.676638] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 642.676819] env[61987]: INFO nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Took 0.05 seconds to destroy the instance on the hypervisor. [ 642.677078] env[61987]: DEBUG oslo.service.loopingcall [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 642.677338] env[61987]: DEBUG nova.compute.manager [-] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 642.677437] env[61987]: DEBUG nova.network.neutron [-] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 642.702613] env[61987]: DEBUG nova.network.neutron [-] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.067603] env[61987]: ERROR nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 26ebdb7d-365e-4aa3-a0ae-35812dd80bbb, please check neutron logs for more information. [ 643.067603] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 643.067603] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 643.067603] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 643.067603] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 643.067603] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 643.067603] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 643.067603] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 643.067603] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 643.067603] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 643.067603] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 643.067603] env[61987]: ERROR nova.compute.manager raise self.value [ 643.067603] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 643.067603] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 643.067603] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 643.067603] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 643.068275] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 643.068275] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 643.068275] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 26ebdb7d-365e-4aa3-a0ae-35812dd80bbb, please check neutron logs for more information. [ 643.068275] env[61987]: ERROR nova.compute.manager [ 643.068275] env[61987]: Traceback (most recent call last): [ 643.068275] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 643.068275] env[61987]: listener.cb(fileno) [ 643.068275] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 643.068275] env[61987]: result = function(*args, **kwargs) [ 643.068275] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 643.068275] env[61987]: return func(*args, **kwargs) [ 643.068275] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 643.068275] env[61987]: raise e [ 643.068275] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 643.068275] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 643.068275] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 643.068275] env[61987]: created_port_ids = self._update_ports_for_instance( [ 643.068275] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 643.068275] env[61987]: with excutils.save_and_reraise_exception(): [ 643.068275] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 643.068275] env[61987]: self.force_reraise() [ 643.068275] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 643.068275] env[61987]: raise self.value [ 643.068275] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 643.068275] env[61987]: updated_port = self._update_port( [ 643.068275] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 643.068275] env[61987]: _ensure_no_port_binding_failure(port) [ 643.068275] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 643.068275] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 643.069076] env[61987]: nova.exception.PortBindingFailed: Binding failed for port 26ebdb7d-365e-4aa3-a0ae-35812dd80bbb, please check neutron logs for more information. [ 643.069076] env[61987]: Removing descriptor: 18 [ 643.158305] env[61987]: DEBUG nova.network.neutron [req-c50229fa-5ca3-4ae8-b3c0-13ecaa17963a req-a736b411-0ad1-4741-a02b-90fa5800c7a0 service nova] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.168017] env[61987]: DEBUG nova.scheduler.client.report [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 643.169656] env[61987]: DEBUG nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 643.200654] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 643.200920] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 643.201101] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 643.201335] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 643.201485] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 643.201631] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 643.201835] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 643.201992] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 643.202171] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 643.202340] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 643.202494] env[61987]: DEBUG nova.virt.hardware [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 643.203689] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3fb6d86-0009-4f84-b549-5cb9f1dc195d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.208238] env[61987]: DEBUG nova.network.neutron [-] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.215025] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7f9bc44-d10d-4403-b856-20c74b46d050 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.230478] env[61987]: ERROR nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 26ebdb7d-365e-4aa3-a0ae-35812dd80bbb, please check neutron logs for more information. [ 643.230478] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Traceback (most recent call last): [ 643.230478] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 643.230478] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] yield resources [ 643.230478] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 643.230478] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] self.driver.spawn(context, instance, image_meta, [ 643.230478] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 643.230478] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 643.230478] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 643.230478] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] vm_ref = self.build_virtual_machine(instance, [ 643.230478] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 643.230819] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] vif_infos = vmwarevif.get_vif_info(self._session, [ 643.230819] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 643.230819] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] for vif in network_info: [ 643.230819] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 643.230819] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] return self._sync_wrapper(fn, *args, **kwargs) [ 643.230819] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 643.230819] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] self.wait() [ 643.230819] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 643.230819] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] self[:] = self._gt.wait() [ 643.230819] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 643.230819] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] return self._exit_event.wait() [ 643.230819] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 643.230819] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] current.throw(*self._exc) [ 643.231219] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 643.231219] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] result = function(*args, **kwargs) [ 643.231219] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 643.231219] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] return func(*args, **kwargs) [ 643.231219] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 643.231219] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] raise e [ 643.231219] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 643.231219] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] nwinfo = self.network_api.allocate_for_instance( [ 643.231219] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 643.231219] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] created_port_ids = self._update_ports_for_instance( [ 643.231219] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 643.231219] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] with excutils.save_and_reraise_exception(): [ 643.231219] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 643.231605] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] self.force_reraise() [ 643.231605] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 643.231605] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] raise self.value [ 643.231605] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 643.231605] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] updated_port = self._update_port( [ 643.231605] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 643.231605] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] _ensure_no_port_binding_failure(port) [ 643.231605] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 643.231605] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] raise exception.PortBindingFailed(port_id=port['id']) [ 643.231605] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] nova.exception.PortBindingFailed: Binding failed for port 26ebdb7d-365e-4aa3-a0ae-35812dd80bbb, please check neutron logs for more information. [ 643.231605] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] [ 643.231605] env[61987]: INFO nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Terminating instance [ 643.251780] env[61987]: DEBUG nova.network.neutron [req-c50229fa-5ca3-4ae8-b3c0-13ecaa17963a req-a736b411-0ad1-4741-a02b-90fa5800c7a0 service nova] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.672903] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.531s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.673533] env[61987]: DEBUG nova.compute.manager [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 643.676060] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.185s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.677437] env[61987]: INFO nova.compute.claims [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 643.710538] env[61987]: INFO nova.compute.manager [-] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Took 1.03 seconds to deallocate network for instance. [ 643.713698] env[61987]: DEBUG nova.compute.claims [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 643.713921] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.734985] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "refresh_cache-65ddff8e-971c-4485-b658-4611aea1176c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.735165] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquired lock "refresh_cache-65ddff8e-971c-4485-b658-4611aea1176c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.735340] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 643.754067] env[61987]: DEBUG oslo_concurrency.lockutils [req-c50229fa-5ca3-4ae8-b3c0-13ecaa17963a req-a736b411-0ad1-4741-a02b-90fa5800c7a0 service nova] Releasing lock "refresh_cache-699c853a-a88f-4890-b712-00f00a6c9838" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.754321] env[61987]: DEBUG nova.compute.manager [req-c50229fa-5ca3-4ae8-b3c0-13ecaa17963a req-a736b411-0ad1-4741-a02b-90fa5800c7a0 service nova] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Received event network-vif-deleted-59570761-d84d-4350-9f6c-17502c6e58ff {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 644.005719] env[61987]: DEBUG nova.compute.manager [req-5368e2dc-82e7-4445-a284-6c23cb9c1c76 req-b7d51628-57a5-4070-83d6-fc49d12b8079 service nova] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Received event network-changed-26ebdb7d-365e-4aa3-a0ae-35812dd80bbb {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 644.005915] env[61987]: DEBUG nova.compute.manager [req-5368e2dc-82e7-4445-a284-6c23cb9c1c76 req-b7d51628-57a5-4070-83d6-fc49d12b8079 service nova] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Refreshing instance network info cache due to event network-changed-26ebdb7d-365e-4aa3-a0ae-35812dd80bbb. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 644.006122] env[61987]: DEBUG oslo_concurrency.lockutils [req-5368e2dc-82e7-4445-a284-6c23cb9c1c76 req-b7d51628-57a5-4070-83d6-fc49d12b8079 service nova] Acquiring lock "refresh_cache-65ddff8e-971c-4485-b658-4611aea1176c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.184381] env[61987]: DEBUG nova.compute.utils [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 644.186896] env[61987]: DEBUG nova.compute.manager [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 644.186896] env[61987]: DEBUG nova.network.neutron [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 644.243146] env[61987]: DEBUG nova.policy [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cc4d69a44feb49c18c22505a8f4e84c9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8c17c124c6c74011894c896a918534d1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 644.258829] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 644.395527] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.678903] env[61987]: DEBUG nova.network.neutron [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Successfully created port: bf22278e-072b-41be-bcd4-ec58e771be7d {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 644.688955] env[61987]: DEBUG nova.compute.manager [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 644.898604] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Releasing lock "refresh_cache-65ddff8e-971c-4485-b658-4611aea1176c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.899055] env[61987]: DEBUG nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 644.899265] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 644.899618] env[61987]: DEBUG oslo_concurrency.lockutils [req-5368e2dc-82e7-4445-a284-6c23cb9c1c76 req-b7d51628-57a5-4070-83d6-fc49d12b8079 service nova] Acquired lock "refresh_cache-65ddff8e-971c-4485-b658-4611aea1176c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.899792] env[61987]: DEBUG nova.network.neutron [req-5368e2dc-82e7-4445-a284-6c23cb9c1c76 req-b7d51628-57a5-4070-83d6-fc49d12b8079 service nova] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Refreshing network info cache for port 26ebdb7d-365e-4aa3-a0ae-35812dd80bbb {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 644.900847] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4bbe666c-7a51-46b4-9793-756fc7bd2160 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.911770] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a239eaa-b46c-4238-b55c-26e21b38fa17 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.939168] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 65ddff8e-971c-4485-b658-4611aea1176c could not be found. [ 644.939168] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 644.939168] env[61987]: INFO nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 644.939168] env[61987]: DEBUG oslo.service.loopingcall [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 644.941853] env[61987]: DEBUG nova.compute.manager [-] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 644.941967] env[61987]: DEBUG nova.network.neutron [-] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 644.963260] env[61987]: DEBUG nova.network.neutron [-] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 645.167248] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b94e4eec-7253-4c6b-9664-d50d8d01051b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.174258] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ca724ee-0618-43df-ace4-78b9548f8469 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.212204] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3e454e5-f055-4bb7-9093-3ca5fc2d7c31 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.219836] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddaec16e-bdac-4aa4-b45d-4110c083c5a1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.233486] env[61987]: DEBUG nova.compute.provider_tree [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 645.438599] env[61987]: DEBUG nova.network.neutron [req-5368e2dc-82e7-4445-a284-6c23cb9c1c76 req-b7d51628-57a5-4070-83d6-fc49d12b8079 service nova] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 645.468578] env[61987]: DEBUG nova.network.neutron [-] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.578730] env[61987]: DEBUG nova.network.neutron [req-5368e2dc-82e7-4445-a284-6c23cb9c1c76 req-b7d51628-57a5-4070-83d6-fc49d12b8079 service nova] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.713753] env[61987]: DEBUG nova.compute.manager [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 645.738863] env[61987]: DEBUG nova.scheduler.client.report [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 645.751144] env[61987]: DEBUG nova.virt.hardware [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 645.755023] env[61987]: DEBUG nova.virt.hardware [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 645.755023] env[61987]: DEBUG nova.virt.hardware [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 645.755023] env[61987]: DEBUG nova.virt.hardware [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 645.755023] env[61987]: DEBUG nova.virt.hardware [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 645.755023] env[61987]: DEBUG nova.virt.hardware [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 645.755220] env[61987]: DEBUG nova.virt.hardware [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 645.755220] env[61987]: DEBUG nova.virt.hardware [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 645.755220] env[61987]: DEBUG nova.virt.hardware [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 645.755220] env[61987]: DEBUG nova.virt.hardware [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 645.755220] env[61987]: DEBUG nova.virt.hardware [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 645.755365] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ab33e47-9f62-4773-9025-54a764a1aebb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.765243] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79f01019-25b3-4e39-a63c-be6fa6131e2a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.786743] env[61987]: ERROR nova.compute.manager [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bf22278e-072b-41be-bcd4-ec58e771be7d, please check neutron logs for more information. [ 645.786743] env[61987]: ERROR nova.compute.manager Traceback (most recent call last): [ 645.786743] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 645.786743] env[61987]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 645.786743] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 645.786743] env[61987]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 645.786743] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 645.786743] env[61987]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 645.786743] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.786743] env[61987]: ERROR nova.compute.manager self.force_reraise() [ 645.786743] env[61987]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.786743] env[61987]: ERROR nova.compute.manager raise self.value [ 645.786743] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 645.786743] env[61987]: ERROR nova.compute.manager updated_port = self._update_port( [ 645.786743] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.786743] env[61987]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 645.787212] env[61987]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.787212] env[61987]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 645.787212] env[61987]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bf22278e-072b-41be-bcd4-ec58e771be7d, please check neutron logs for more information. [ 645.787212] env[61987]: ERROR nova.compute.manager [ 645.787212] env[61987]: Traceback (most recent call last): [ 645.787212] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 645.787212] env[61987]: listener.cb(fileno) [ 645.787212] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.787212] env[61987]: result = function(*args, **kwargs) [ 645.787212] env[61987]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 645.787212] env[61987]: return func(*args, **kwargs) [ 645.787212] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 645.787212] env[61987]: raise e [ 645.787212] env[61987]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 645.787212] env[61987]: nwinfo = self.network_api.allocate_for_instance( [ 645.787212] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 645.787212] env[61987]: created_port_ids = self._update_ports_for_instance( [ 645.787212] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 645.787212] env[61987]: with excutils.save_and_reraise_exception(): [ 645.787212] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.787212] env[61987]: self.force_reraise() [ 645.787212] env[61987]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.787212] env[61987]: raise self.value [ 645.787212] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 645.787212] env[61987]: updated_port = self._update_port( [ 645.787212] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.787212] env[61987]: _ensure_no_port_binding_failure(port) [ 645.787212] env[61987]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.787212] env[61987]: raise exception.PortBindingFailed(port_id=port['id']) [ 645.788085] env[61987]: nova.exception.PortBindingFailed: Binding failed for port bf22278e-072b-41be-bcd4-ec58e771be7d, please check neutron logs for more information. [ 645.788085] env[61987]: Removing descriptor: 18 [ 645.788085] env[61987]: ERROR nova.compute.manager [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bf22278e-072b-41be-bcd4-ec58e771be7d, please check neutron logs for more information. [ 645.788085] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Traceback (most recent call last): [ 645.788085] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 645.788085] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] yield resources [ 645.788085] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 645.788085] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] self.driver.spawn(context, instance, image_meta, [ 645.788085] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 645.788085] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 645.788085] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 645.788085] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] vm_ref = self.build_virtual_machine(instance, [ 645.788460] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 645.788460] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] vif_infos = vmwarevif.get_vif_info(self._session, [ 645.788460] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 645.788460] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] for vif in network_info: [ 645.788460] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 645.788460] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] return self._sync_wrapper(fn, *args, **kwargs) [ 645.788460] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 645.788460] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] self.wait() [ 645.788460] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 645.788460] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] self[:] = self._gt.wait() [ 645.788460] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 645.788460] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] return self._exit_event.wait() [ 645.788460] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 645.788822] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] result = hub.switch() [ 645.788822] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 645.788822] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] return self.greenlet.switch() [ 645.788822] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.788822] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] result = function(*args, **kwargs) [ 645.788822] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 645.788822] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] return func(*args, **kwargs) [ 645.788822] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 645.788822] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] raise e [ 645.788822] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 645.788822] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] nwinfo = self.network_api.allocate_for_instance( [ 645.788822] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 645.788822] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] created_port_ids = self._update_ports_for_instance( [ 645.789203] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 645.789203] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] with excutils.save_and_reraise_exception(): [ 645.789203] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.789203] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] self.force_reraise() [ 645.789203] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.789203] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] raise self.value [ 645.789203] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 645.789203] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] updated_port = self._update_port( [ 645.789203] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.789203] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] _ensure_no_port_binding_failure(port) [ 645.789203] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.789203] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] raise exception.PortBindingFailed(port_id=port['id']) [ 645.789531] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] nova.exception.PortBindingFailed: Binding failed for port bf22278e-072b-41be-bcd4-ec58e771be7d, please check neutron logs for more information. [ 645.789531] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] [ 645.789531] env[61987]: INFO nova.compute.manager [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Terminating instance [ 645.971771] env[61987]: INFO nova.compute.manager [-] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Took 1.03 seconds to deallocate network for instance. [ 645.974363] env[61987]: DEBUG nova.compute.claims [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 645.974560] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.082678] env[61987]: DEBUG oslo_concurrency.lockutils [req-5368e2dc-82e7-4445-a284-6c23cb9c1c76 req-b7d51628-57a5-4070-83d6-fc49d12b8079 service nova] Releasing lock "refresh_cache-65ddff8e-971c-4485-b658-4611aea1176c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.082678] env[61987]: DEBUG nova.compute.manager [req-5368e2dc-82e7-4445-a284-6c23cb9c1c76 req-b7d51628-57a5-4070-83d6-fc49d12b8079 service nova] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Received event network-vif-deleted-26ebdb7d-365e-4aa3-a0ae-35812dd80bbb {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 646.082678] env[61987]: DEBUG nova.compute.manager [req-82928cc6-fcab-48f9-a72d-a0ff3f1e44d1 req-61445d42-4bc1-4d0f-8f46-a33b40c926d5 service nova] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Received event network-changed-bf22278e-072b-41be-bcd4-ec58e771be7d {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 646.082814] env[61987]: DEBUG nova.compute.manager [req-82928cc6-fcab-48f9-a72d-a0ff3f1e44d1 req-61445d42-4bc1-4d0f-8f46-a33b40c926d5 service nova] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Refreshing instance network info cache due to event network-changed-bf22278e-072b-41be-bcd4-ec58e771be7d. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 646.083017] env[61987]: DEBUG oslo_concurrency.lockutils [req-82928cc6-fcab-48f9-a72d-a0ff3f1e44d1 req-61445d42-4bc1-4d0f-8f46-a33b40c926d5 service nova] Acquiring lock "refresh_cache-f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.083161] env[61987]: DEBUG oslo_concurrency.lockutils [req-82928cc6-fcab-48f9-a72d-a0ff3f1e44d1 req-61445d42-4bc1-4d0f-8f46-a33b40c926d5 service nova] Acquired lock "refresh_cache-f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.083388] env[61987]: DEBUG nova.network.neutron [req-82928cc6-fcab-48f9-a72d-a0ff3f1e44d1 req-61445d42-4bc1-4d0f-8f46-a33b40c926d5 service nova] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Refreshing network info cache for port bf22278e-072b-41be-bcd4-ec58e771be7d {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 646.246839] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.571s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.247458] env[61987]: DEBUG nova.compute.manager [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 646.250786] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.798s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.251386] env[61987]: DEBUG nova.objects.instance [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Lazy-loading 'resources' on Instance uuid e6c56d86-e428-49fb-9944-293ad5e5948c {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 646.292325] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Acquiring lock "refresh_cache-f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.601725] env[61987]: DEBUG nova.network.neutron [req-82928cc6-fcab-48f9-a72d-a0ff3f1e44d1 req-61445d42-4bc1-4d0f-8f46-a33b40c926d5 service nova] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 646.659952] env[61987]: DEBUG nova.network.neutron [req-82928cc6-fcab-48f9-a72d-a0ff3f1e44d1 req-61445d42-4bc1-4d0f-8f46-a33b40c926d5 service nova] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.758543] env[61987]: DEBUG nova.compute.utils [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 646.759920] env[61987]: DEBUG nova.compute.manager [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 646.760101] env[61987]: DEBUG nova.network.neutron [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 646.816878] env[61987]: DEBUG nova.policy [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fadc61080dcb4a229397575b17c111c8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a082e4efc2ed4c7abfe86b1336fa6b2d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 647.164967] env[61987]: DEBUG oslo_concurrency.lockutils [req-82928cc6-fcab-48f9-a72d-a0ff3f1e44d1 req-61445d42-4bc1-4d0f-8f46-a33b40c926d5 service nova] Releasing lock "refresh_cache-f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.164967] env[61987]: DEBUG nova.compute.manager [req-82928cc6-fcab-48f9-a72d-a0ff3f1e44d1 req-61445d42-4bc1-4d0f-8f46-a33b40c926d5 service nova] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Received event network-vif-deleted-bf22278e-072b-41be-bcd4-ec58e771be7d {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 647.164967] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Acquired lock "refresh_cache-f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.164967] env[61987]: DEBUG nova.network.neutron [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 647.169410] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af88032c-0a64-4c7e-ba8c-2afd1ec19fcb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.178510] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da86955d-af15-48a7-a354-155fc4d697c0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.212296] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8594d34b-5e64-4afa-99d3-00f1242abf04 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.220864] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e31bcdd-71f9-4b7d-b088-d87009a65f22 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.235016] env[61987]: DEBUG nova.compute.provider_tree [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.236561] env[61987]: DEBUG nova.network.neutron [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Successfully created port: 90e59ebc-8840-4623-9c9b-2730ddf7f708 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 647.266135] env[61987]: DEBUG nova.compute.manager [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 647.682746] env[61987]: DEBUG nova.network.neutron [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 647.731795] env[61987]: DEBUG nova.network.neutron [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.740234] env[61987]: DEBUG nova.scheduler.client.report [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 648.234697] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Releasing lock "refresh_cache-f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.235160] env[61987]: DEBUG nova.compute.manager [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 648.237021] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 648.237021] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2aee8ce9-1e27-45ee-b674-093776e2734d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.245058] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b238da7b-0521-492d-bad2-9c266bbf9a5e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.255829] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.005s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.257812] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 20.925s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.257985] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.258154] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61987) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 648.258415] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.309s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.262678] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b59770-4704-4986-bb2d-2e27c6d18c47 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.275363] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f92e4785-f8e1-4aab-82ac-adfaeca094c8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.279837] env[61987]: DEBUG nova.compute.manager [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 648.281940] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0 could not be found. [ 648.282154] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 648.282335] env[61987]: INFO nova.compute.manager [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Took 0.05 seconds to destroy the instance on the hypervisor. [ 648.282601] env[61987]: DEBUG oslo.service.loopingcall [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 648.283030] env[61987]: DEBUG nova.compute.manager [-] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 648.283138] env[61987]: DEBUG nova.network.neutron [-] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 648.297024] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-113cea14-3798-40a8-9723-cf415707dcc4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.299039] env[61987]: DEBUG nova.network.neutron [-] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 648.306028] env[61987]: INFO nova.scheduler.client.report [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Deleted allocations for instance e6c56d86-e428-49fb-9944-293ad5e5948c [ 648.311340] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-044be69a-7c88-4c52-8653-eb1462ef6585 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.345682] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181568MB free_disk=180GB free_vcpus=48 pci_devices=None {{(pid=61987) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 648.345992] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.348391] env[61987]: DEBUG nova.virt.hardware [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 648.348638] env[61987]: DEBUG nova.virt.hardware [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 648.348800] env[61987]: DEBUG nova.virt.hardware [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 648.348986] env[61987]: DEBUG nova.virt.hardware [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 648.349147] env[61987]: DEBUG nova.virt.hardware [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 648.349298] env[61987]: DEBUG nova.virt.hardware [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 648.349504] env[61987]: DEBUG nova.virt.hardware [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 648.349692] env[61987]: DEBUG nova.virt.hardware [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 648.349851] env[61987]: DEBUG nova.virt.hardware [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 648.350025] env[61987]: DEBUG nova.virt.hardware [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 648.350204] env[61987]: DEBUG nova.virt.hardware [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 648.351033] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d3345f-d510-4a6c-a039-2c9a55d5db38 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.359477] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab6d5548-c89a-41d3-9000-829c147f15c7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.801848] env[61987]: DEBUG nova.network.neutron [-] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.820476] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88b8cdf2-f10c-4e06-93f8-cfbbac9770d7 tempest-ServersAaction247Test-990889782 tempest-ServersAaction247Test-990889782-project-member] Lock "e6c56d86-e428-49fb-9944-293ad5e5948c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.722s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.869324] env[61987]: DEBUG nova.compute.manager [req-ea0e714a-9be1-4aab-a529-ddd256af5ca0 req-af1b990a-298b-44f7-98fb-57eadace20fc service nova] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Received event network-vif-plugged-90e59ebc-8840-4623-9c9b-2730ddf7f708 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 648.869543] env[61987]: DEBUG oslo_concurrency.lockutils [req-ea0e714a-9be1-4aab-a529-ddd256af5ca0 req-af1b990a-298b-44f7-98fb-57eadace20fc service nova] Acquiring lock "6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.869729] env[61987]: DEBUG oslo_concurrency.lockutils [req-ea0e714a-9be1-4aab-a529-ddd256af5ca0 req-af1b990a-298b-44f7-98fb-57eadace20fc service nova] Lock "6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.869888] env[61987]: DEBUG oslo_concurrency.lockutils [req-ea0e714a-9be1-4aab-a529-ddd256af5ca0 req-af1b990a-298b-44f7-98fb-57eadace20fc service nova] Lock "6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.870061] env[61987]: DEBUG nova.compute.manager [req-ea0e714a-9be1-4aab-a529-ddd256af5ca0 req-af1b990a-298b-44f7-98fb-57eadace20fc service nova] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] No waiting events found dispatching network-vif-plugged-90e59ebc-8840-4623-9c9b-2730ddf7f708 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 648.870226] env[61987]: WARNING nova.compute.manager [req-ea0e714a-9be1-4aab-a529-ddd256af5ca0 req-af1b990a-298b-44f7-98fb-57eadace20fc service nova] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Received unexpected event network-vif-plugged-90e59ebc-8840-4623-9c9b-2730ddf7f708 for instance with vm_state building and task_state spawning. [ 648.963583] env[61987]: DEBUG nova.network.neutron [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Successfully updated port: 90e59ebc-8840-4623-9c9b-2730ddf7f708 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 649.206560] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c7ccf6a-4557-4a85-bed2-745308a2223b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.213959] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37795117-7eb4-40f5-9573-83a46fd974be {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.244035] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cd7d454-2a15-4852-b4f3-be13c45b676f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.251543] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a87211-3796-4785-a3a9-baf3440260ba {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.265680] env[61987]: DEBUG nova.compute.provider_tree [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.305629] env[61987]: INFO nova.compute.manager [-] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Took 1.02 seconds to deallocate network for instance. [ 649.308081] env[61987]: DEBUG nova.compute.claims [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Aborting claim: {{(pid=61987) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 649.308289] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.467122] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Acquiring lock "refresh_cache-6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.467968] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Acquired lock "refresh_cache-6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.467968] env[61987]: DEBUG nova.network.neutron [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 649.769399] env[61987]: DEBUG nova.scheduler.client.report [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 650.002056] env[61987]: DEBUG nova.network.neutron [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 650.181788] env[61987]: DEBUG nova.network.neutron [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Updating instance_info_cache with network_info: [{"id": "90e59ebc-8840-4623-9c9b-2730ddf7f708", "address": "fa:16:3e:f4:e7:34", "network": {"id": "8a9770a4-fed3-4232-ac78-8a1fc2223229", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.70", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "705320a8b6a549d0b197ee1a35404d52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bafe8721-91d4-4127-b215-d9e8e27947dc", "external-id": "nsx-vlan-transportzone-680", "segmentation_id": 680, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90e59ebc-88", "ovs_interfaceid": "90e59ebc-8840-4623-9c9b-2730ddf7f708", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.275527] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.017s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.276296] env[61987]: ERROR nova.compute.manager [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cf948a39-baf8-4fe5-b691-cea4f5698e8f, please check neutron logs for more information. [ 650.276296] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Traceback (most recent call last): [ 650.276296] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 650.276296] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] self.driver.spawn(context, instance, image_meta, [ 650.276296] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 650.276296] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 650.276296] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 650.276296] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] vm_ref = self.build_virtual_machine(instance, [ 650.276296] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 650.276296] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] vif_infos = vmwarevif.get_vif_info(self._session, [ 650.276296] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 650.276663] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] for vif in network_info: [ 650.276663] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 650.276663] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] return self._sync_wrapper(fn, *args, **kwargs) [ 650.276663] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 650.276663] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] self.wait() [ 650.276663] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 650.276663] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] self[:] = self._gt.wait() [ 650.276663] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 650.276663] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] return self._exit_event.wait() [ 650.276663] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 650.276663] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] result = hub.switch() [ 650.276663] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 650.276663] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] return self.greenlet.switch() [ 650.277010] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.277010] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] result = function(*args, **kwargs) [ 650.277010] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 650.277010] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] return func(*args, **kwargs) [ 650.277010] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 650.277010] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] raise e [ 650.277010] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 650.277010] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] nwinfo = self.network_api.allocate_for_instance( [ 650.277010] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 650.277010] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] created_port_ids = self._update_ports_for_instance( [ 650.277010] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 650.277010] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] with excutils.save_and_reraise_exception(): [ 650.277010] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.277373] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] self.force_reraise() [ 650.277373] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.277373] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] raise self.value [ 650.277373] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 650.277373] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] updated_port = self._update_port( [ 650.277373] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.277373] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] _ensure_no_port_binding_failure(port) [ 650.277373] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.277373] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] raise exception.PortBindingFailed(port_id=port['id']) [ 650.277373] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] nova.exception.PortBindingFailed: Binding failed for port cf948a39-baf8-4fe5-b691-cea4f5698e8f, please check neutron logs for more information. [ 650.277373] env[61987]: ERROR nova.compute.manager [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] [ 650.277690] env[61987]: DEBUG nova.compute.utils [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Binding failed for port cf948a39-baf8-4fe5-b691-cea4f5698e8f, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 650.278517] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.503s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.281534] env[61987]: DEBUG nova.compute.manager [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Build of instance 40fefd68-3d6f-4451-997f-e8fa49847dea was re-scheduled: Binding failed for port cf948a39-baf8-4fe5-b691-cea4f5698e8f, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 650.282111] env[61987]: DEBUG nova.compute.manager [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 650.282338] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "refresh_cache-40fefd68-3d6f-4451-997f-e8fa49847dea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.282479] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired lock "refresh_cache-40fefd68-3d6f-4451-997f-e8fa49847dea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.282635] env[61987]: DEBUG nova.network.neutron [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 650.684229] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Releasing lock "refresh_cache-6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.684576] env[61987]: DEBUG nova.compute.manager [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Instance network_info: |[{"id": "90e59ebc-8840-4623-9c9b-2730ddf7f708", "address": "fa:16:3e:f4:e7:34", "network": {"id": "8a9770a4-fed3-4232-ac78-8a1fc2223229", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.70", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "705320a8b6a549d0b197ee1a35404d52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bafe8721-91d4-4127-b215-d9e8e27947dc", "external-id": "nsx-vlan-transportzone-680", "segmentation_id": 680, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90e59ebc-88", "ovs_interfaceid": "90e59ebc-8840-4623-9c9b-2730ddf7f708", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 650.685063] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f4:e7:34', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bafe8721-91d4-4127-b215-d9e8e27947dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '90e59ebc-8840-4623-9c9b-2730ddf7f708', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 650.692804] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Creating folder: Project (a082e4efc2ed4c7abfe86b1336fa6b2d). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 650.693111] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-83931b32-dd90-4686-b3c2-3f79720d556a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.704312] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Created folder: Project (a082e4efc2ed4c7abfe86b1336fa6b2d) in parent group-v234219. [ 650.704499] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Creating folder: Instances. Parent ref: group-v234230. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 650.704722] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-29bb865a-6129-43de-9179-a0b0a8893bb7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.712859] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Created folder: Instances in parent group-v234230. [ 650.713099] env[61987]: DEBUG oslo.service.loopingcall [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 650.713305] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 650.713503] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-26a240f7-81d2-44fb-b5f8-0fb545f8de15 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.732931] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 650.732931] env[61987]: value = "task-1061270" [ 650.732931] env[61987]: _type = "Task" [ 650.732931] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.740038] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061270, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.818984] env[61987]: DEBUG nova.network.neutron [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 650.943579] env[61987]: DEBUG nova.network.neutron [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.032750] env[61987]: DEBUG nova.compute.manager [req-823d4b97-e5cf-44a5-a7a9-4aee8efff423 req-a96ed87b-3c67-410f-94bc-b52c6516484b service nova] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Received event network-changed-90e59ebc-8840-4623-9c9b-2730ddf7f708 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 651.032952] env[61987]: DEBUG nova.compute.manager [req-823d4b97-e5cf-44a5-a7a9-4aee8efff423 req-a96ed87b-3c67-410f-94bc-b52c6516484b service nova] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Refreshing instance network info cache due to event network-changed-90e59ebc-8840-4623-9c9b-2730ddf7f708. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 651.033188] env[61987]: DEBUG oslo_concurrency.lockutils [req-823d4b97-e5cf-44a5-a7a9-4aee8efff423 req-a96ed87b-3c67-410f-94bc-b52c6516484b service nova] Acquiring lock "refresh_cache-6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.033382] env[61987]: DEBUG oslo_concurrency.lockutils [req-823d4b97-e5cf-44a5-a7a9-4aee8efff423 req-a96ed87b-3c67-410f-94bc-b52c6516484b service nova] Acquired lock "refresh_cache-6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.033557] env[61987]: DEBUG nova.network.neutron [req-823d4b97-e5cf-44a5-a7a9-4aee8efff423 req-a96ed87b-3c67-410f-94bc-b52c6516484b service nova] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Refreshing network info cache for port 90e59ebc-8840-4623-9c9b-2730ddf7f708 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 651.175436] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e034d3a-5661-423e-995b-9aa77a7f6459 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.183306] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e75acf16-c451-4496-a90f-a0667c6dceda {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.212873] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19890564-83d8-4ca8-ba3e-7093492fc186 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.220753] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dae3037-3ec2-4e99-8815-abcc41f07712 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.233865] env[61987]: DEBUG nova.compute.provider_tree [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.242400] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061270, 'name': CreateVM_Task, 'duration_secs': 0.327344} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.243124] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 651.250491] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.250665] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.250969] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 651.251221] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32e9043d-c06a-43d3-8362-77f5f47b640f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.257262] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Waiting for the task: (returnval){ [ 651.257262] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52dee986-156a-a806-a5a5-ea011e9de6e3" [ 651.257262] env[61987]: _type = "Task" [ 651.257262] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.264836] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52dee986-156a-a806-a5a5-ea011e9de6e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.448834] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Releasing lock "refresh_cache-40fefd68-3d6f-4451-997f-e8fa49847dea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.449140] env[61987]: DEBUG nova.compute.manager [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 651.449263] env[61987]: DEBUG nova.compute.manager [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 651.449432] env[61987]: DEBUG nova.network.neutron [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 651.465666] env[61987]: DEBUG nova.network.neutron [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 651.739756] env[61987]: DEBUG nova.scheduler.client.report [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 651.747913] env[61987]: DEBUG nova.network.neutron [req-823d4b97-e5cf-44a5-a7a9-4aee8efff423 req-a96ed87b-3c67-410f-94bc-b52c6516484b service nova] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Updated VIF entry in instance network info cache for port 90e59ebc-8840-4623-9c9b-2730ddf7f708. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 651.748352] env[61987]: DEBUG nova.network.neutron [req-823d4b97-e5cf-44a5-a7a9-4aee8efff423 req-a96ed87b-3c67-410f-94bc-b52c6516484b service nova] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Updating instance_info_cache with network_info: [{"id": "90e59ebc-8840-4623-9c9b-2730ddf7f708", "address": "fa:16:3e:f4:e7:34", "network": {"id": "8a9770a4-fed3-4232-ac78-8a1fc2223229", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.70", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "705320a8b6a549d0b197ee1a35404d52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bafe8721-91d4-4127-b215-d9e8e27947dc", "external-id": "nsx-vlan-transportzone-680", "segmentation_id": 680, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90e59ebc-88", "ovs_interfaceid": "90e59ebc-8840-4623-9c9b-2730ddf7f708", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.769866] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52dee986-156a-a806-a5a5-ea011e9de6e3, 'name': SearchDatastore_Task, 'duration_secs': 0.010642} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.770028] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.770259] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 651.770562] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.770788] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.771020] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 651.771568] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-74345fe9-ac4d-473d-95ff-e8d246d70d71 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.780399] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 651.780709] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 651.781492] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d2db99d-dda6-4cdc-a0f1-a75183e62f81 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.787525] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Waiting for the task: (returnval){ [ 651.787525] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]522e5342-c4e8-16bf-388f-6010e4ecf82b" [ 651.787525] env[61987]: _type = "Task" [ 651.787525] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.797042] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]522e5342-c4e8-16bf-388f-6010e4ecf82b, 'name': SearchDatastore_Task, 'duration_secs': 0.007475} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.797810] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26ba1972-9e94-4ec9-b7d8-26d35e8eb936 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.802661] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Waiting for the task: (returnval){ [ 651.802661] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52982f2f-daf6-6b5e-065c-55aef00bea8a" [ 651.802661] env[61987]: _type = "Task" [ 651.802661] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.810983] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52982f2f-daf6-6b5e-065c-55aef00bea8a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.967987] env[61987]: DEBUG nova.network.neutron [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.250801] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.972s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.251484] env[61987]: ERROR nova.compute.manager [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1c4aa77c-6462-4435-bda8-10dc64c8e052, please check neutron logs for more information. [ 652.251484] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Traceback (most recent call last): [ 652.251484] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 652.251484] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] self.driver.spawn(context, instance, image_meta, [ 652.251484] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 652.251484] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] self._vmops.spawn(context, instance, image_meta, injected_files, [ 652.251484] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 652.251484] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] vm_ref = self.build_virtual_machine(instance, [ 652.251484] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 652.251484] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] vif_infos = vmwarevif.get_vif_info(self._session, [ 652.251484] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 652.251818] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] for vif in network_info: [ 652.251818] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 652.251818] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] return self._sync_wrapper(fn, *args, **kwargs) [ 652.251818] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 652.251818] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] self.wait() [ 652.251818] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 652.251818] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] self[:] = self._gt.wait() [ 652.251818] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 652.251818] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] return self._exit_event.wait() [ 652.251818] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 652.251818] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] result = hub.switch() [ 652.251818] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 652.251818] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] return self.greenlet.switch() [ 652.252186] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.252186] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] result = function(*args, **kwargs) [ 652.252186] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 652.252186] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] return func(*args, **kwargs) [ 652.252186] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 652.252186] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] raise e [ 652.252186] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 652.252186] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] nwinfo = self.network_api.allocate_for_instance( [ 652.252186] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.252186] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] created_port_ids = self._update_ports_for_instance( [ 652.252186] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.252186] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] with excutils.save_and_reraise_exception(): [ 652.252186] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.252546] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] self.force_reraise() [ 652.252546] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.252546] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] raise self.value [ 652.252546] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.252546] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] updated_port = self._update_port( [ 652.252546] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.252546] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] _ensure_no_port_binding_failure(port) [ 652.252546] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.252546] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] raise exception.PortBindingFailed(port_id=port['id']) [ 652.252546] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] nova.exception.PortBindingFailed: Binding failed for port 1c4aa77c-6462-4435-bda8-10dc64c8e052, please check neutron logs for more information. [ 652.252546] env[61987]: ERROR nova.compute.manager [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] [ 652.252858] env[61987]: DEBUG nova.compute.utils [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Binding failed for port 1c4aa77c-6462-4435-bda8-10dc64c8e052, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 652.253563] env[61987]: DEBUG oslo_concurrency.lockutils [req-823d4b97-e5cf-44a5-a7a9-4aee8efff423 req-a96ed87b-3c67-410f-94bc-b52c6516484b service nova] Releasing lock "refresh_cache-6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.253943] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.855s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.256695] env[61987]: DEBUG nova.compute.manager [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Build of instance c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec was re-scheduled: Binding failed for port 1c4aa77c-6462-4435-bda8-10dc64c8e052, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 652.257127] env[61987]: DEBUG nova.compute.manager [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 652.257345] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Acquiring lock "refresh_cache-c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.257493] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Acquired lock "refresh_cache-c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.257648] env[61987]: DEBUG nova.network.neutron [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 652.312637] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52982f2f-daf6-6b5e-065c-55aef00bea8a, 'name': SearchDatastore_Task, 'duration_secs': 0.008643} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.312874] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.313145] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59/6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 652.313428] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cdc89999-45e3-4bcb-8478-c475d1d07fd2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.321049] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Waiting for the task: (returnval){ [ 652.321049] env[61987]: value = "task-1061271" [ 652.321049] env[61987]: _type = "Task" [ 652.321049] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.328220] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Task: {'id': task-1061271, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.470959] env[61987]: INFO nova.compute.manager [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 40fefd68-3d6f-4451-997f-e8fa49847dea] Took 1.02 seconds to deallocate network for instance. [ 652.778950] env[61987]: DEBUG nova.network.neutron [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 652.835110] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Task: {'id': task-1061271, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.440067} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.835436] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59/6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 652.835649] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 652.835898] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8b1cac67-6e40-496c-8710-1e221d36e85e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.841378] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Waiting for the task: (returnval){ [ 652.841378] env[61987]: value = "task-1061272" [ 652.841378] env[61987]: _type = "Task" [ 652.841378] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.849035] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Task: {'id': task-1061272, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.876751] env[61987]: DEBUG nova.network.neutron [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.115332] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49464722-bc67-4ceb-9e09-faf7cee98486 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.123184] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf02d81-ea34-4341-b2ce-30a90530d20d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.155399] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c052036d-cf32-499e-9ae0-ba7dbc763673 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.162635] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-662df9ba-cc0d-4a07-b20e-0e18b9d67812 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.177235] env[61987]: DEBUG nova.compute.provider_tree [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.350706] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Task: {'id': task-1061272, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063816} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.351030] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 653.351836] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d62c0097-1886-46e5-97b4-11754a904e4f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.374465] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Reconfiguring VM instance instance-0000001f to attach disk [datastore2] 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59/6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 653.374738] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-30beedf3-6736-40ab-9e1d-43d3cc244b26 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.389869] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Releasing lock "refresh_cache-c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.390125] env[61987]: DEBUG nova.compute.manager [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 653.390288] env[61987]: DEBUG nova.compute.manager [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 653.390457] env[61987]: DEBUG nova.network.neutron [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 653.396640] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Waiting for the task: (returnval){ [ 653.396640] env[61987]: value = "task-1061273" [ 653.396640] env[61987]: _type = "Task" [ 653.396640] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.407284] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Task: {'id': task-1061273, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.408062] env[61987]: DEBUG nova.network.neutron [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 653.509876] env[61987]: INFO nova.scheduler.client.report [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Deleted allocations for instance 40fefd68-3d6f-4451-997f-e8fa49847dea [ 653.680806] env[61987]: DEBUG nova.scheduler.client.report [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 653.906791] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Task: {'id': task-1061273, 'name': ReconfigVM_Task, 'duration_secs': 0.283611} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.907145] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Reconfigured VM instance instance-0000001f to attach disk [datastore2] 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59/6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 653.907767] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9fd32b7a-7e6b-45a8-aa12-53af0dabe103 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.910823] env[61987]: DEBUG nova.network.neutron [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.913908] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Waiting for the task: (returnval){ [ 653.913908] env[61987]: value = "task-1061274" [ 653.913908] env[61987]: _type = "Task" [ 653.913908] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.921594] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Task: {'id': task-1061274, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.019101] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5dbafd5-50bc-491e-ab05-0e5c7f6743e2 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "40fefd68-3d6f-4451-997f-e8fa49847dea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.009s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.185508] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.931s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.186212] env[61987]: ERROR nova.compute.manager [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port de0f490b-f7ab-4dfe-83b1-3574487c4556, please check neutron logs for more information. [ 654.186212] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Traceback (most recent call last): [ 654.186212] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 654.186212] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] self.driver.spawn(context, instance, image_meta, [ 654.186212] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 654.186212] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 654.186212] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 654.186212] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] vm_ref = self.build_virtual_machine(instance, [ 654.186212] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 654.186212] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] vif_infos = vmwarevif.get_vif_info(self._session, [ 654.186212] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 654.186575] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] for vif in network_info: [ 654.186575] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 654.186575] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] return self._sync_wrapper(fn, *args, **kwargs) [ 654.186575] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 654.186575] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] self.wait() [ 654.186575] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 654.186575] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] self[:] = self._gt.wait() [ 654.186575] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 654.186575] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] return self._exit_event.wait() [ 654.186575] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 654.186575] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] result = hub.switch() [ 654.186575] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 654.186575] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] return self.greenlet.switch() [ 654.186943] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.186943] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] result = function(*args, **kwargs) [ 654.186943] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 654.186943] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] return func(*args, **kwargs) [ 654.186943] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 654.186943] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] raise e [ 654.186943] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 654.186943] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] nwinfo = self.network_api.allocate_for_instance( [ 654.186943] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 654.186943] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] created_port_ids = self._update_ports_for_instance( [ 654.186943] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 654.186943] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] with excutils.save_and_reraise_exception(): [ 654.186943] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.187325] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] self.force_reraise() [ 654.187325] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.187325] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] raise self.value [ 654.187325] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 654.187325] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] updated_port = self._update_port( [ 654.187325] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.187325] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] _ensure_no_port_binding_failure(port) [ 654.187325] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.187325] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] raise exception.PortBindingFailed(port_id=port['id']) [ 654.187325] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] nova.exception.PortBindingFailed: Binding failed for port de0f490b-f7ab-4dfe-83b1-3574487c4556, please check neutron logs for more information. [ 654.187325] env[61987]: ERROR nova.compute.manager [instance: 967bc997-543f-4b35-968b-c530b82d9db0] [ 654.187630] env[61987]: DEBUG nova.compute.utils [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Binding failed for port de0f490b-f7ab-4dfe-83b1-3574487c4556, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 654.188136] env[61987]: DEBUG oslo_concurrency.lockutils [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.085s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.190913] env[61987]: DEBUG nova.compute.manager [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Build of instance 967bc997-543f-4b35-968b-c530b82d9db0 was re-scheduled: Binding failed for port de0f490b-f7ab-4dfe-83b1-3574487c4556, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 654.191537] env[61987]: DEBUG nova.compute.manager [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 654.191614] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Acquiring lock "refresh_cache-967bc997-543f-4b35-968b-c530b82d9db0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.191739] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Acquired lock "refresh_cache-967bc997-543f-4b35-968b-c530b82d9db0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.191898] env[61987]: DEBUG nova.network.neutron [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 654.416386] env[61987]: INFO nova.compute.manager [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec] Took 1.02 seconds to deallocate network for instance. [ 654.426785] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Task: {'id': task-1061274, 'name': Rename_Task, 'duration_secs': 0.144098} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.427070] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 654.427316] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-80e93fe7-1d3d-473d-930e-e433619d94c1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.433321] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Waiting for the task: (returnval){ [ 654.433321] env[61987]: value = "task-1061275" [ 654.433321] env[61987]: _type = "Task" [ 654.433321] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.441917] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Task: {'id': task-1061275, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.522379] env[61987]: DEBUG nova.compute.manager [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 654.714813] env[61987]: DEBUG nova.network.neutron [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.775236] env[61987]: DEBUG nova.network.neutron [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.945491] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Task: {'id': task-1061275, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.047231] env[61987]: DEBUG oslo_concurrency.lockutils [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.059235] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3bddc3-9532-4439-b972-8b2e59411190 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.066510] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd6ffcb2-ed9c-48a6-bce1-1fce66373121 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.096517] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2f95629-3c5b-4956-81b8-4147a51a95e0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.103512] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e3cf506-def1-411f-b0b6-71e60f312a16 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.116304] env[61987]: DEBUG nova.compute.provider_tree [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.277220] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Releasing lock "refresh_cache-967bc997-543f-4b35-968b-c530b82d9db0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.277477] env[61987]: DEBUG nova.compute.manager [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 655.277643] env[61987]: DEBUG nova.compute.manager [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 655.277811] env[61987]: DEBUG nova.network.neutron [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 655.294806] env[61987]: DEBUG nova.network.neutron [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 655.443894] env[61987]: DEBUG oslo_vmware.api [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Task: {'id': task-1061275, 'name': PowerOnVM_Task, 'duration_secs': 1.000818} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.443995] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 655.444847] env[61987]: INFO nova.compute.manager [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Took 7.16 seconds to spawn the instance on the hypervisor. [ 655.444847] env[61987]: DEBUG nova.compute.manager [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 655.445313] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba16d9e7-7e16-44b1-bfd2-7e6871c0e1b4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.448797] env[61987]: INFO nova.scheduler.client.report [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Deleted allocations for instance c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec [ 655.566348] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.566708] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.619527] env[61987]: DEBUG nova.scheduler.client.report [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 655.797648] env[61987]: DEBUG nova.network.neutron [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.959134] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a19cc01-64ee-4e67-bc36-d829f4ca8dbf tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Lock "c8e5d6c1-44a0-465f-a0dd-03e519f3d5ec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.945s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.973512] env[61987]: INFO nova.compute.manager [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Took 32.51 seconds to build instance. [ 656.124885] env[61987]: DEBUG oslo_concurrency.lockutils [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.937s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.125587] env[61987]: ERROR nova.compute.manager [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3041ccf6-e87e-4e4e-8332-1a04de5361ce, please check neutron logs for more information. [ 656.125587] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Traceback (most recent call last): [ 656.125587] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 656.125587] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] self.driver.spawn(context, instance, image_meta, [ 656.125587] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 656.125587] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.125587] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.125587] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] vm_ref = self.build_virtual_machine(instance, [ 656.125587] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.125587] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.125587] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.125915] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] for vif in network_info: [ 656.125915] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 656.125915] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] return self._sync_wrapper(fn, *args, **kwargs) [ 656.125915] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 656.125915] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] self.wait() [ 656.125915] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 656.125915] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] self[:] = self._gt.wait() [ 656.125915] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.125915] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] return self._exit_event.wait() [ 656.125915] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 656.125915] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] result = hub.switch() [ 656.125915] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 656.125915] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] return self.greenlet.switch() [ 656.126284] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.126284] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] result = function(*args, **kwargs) [ 656.126284] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 656.126284] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] return func(*args, **kwargs) [ 656.126284] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 656.126284] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] raise e [ 656.126284] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 656.126284] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] nwinfo = self.network_api.allocate_for_instance( [ 656.126284] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 656.126284] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] created_port_ids = self._update_ports_for_instance( [ 656.126284] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 656.126284] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] with excutils.save_and_reraise_exception(): [ 656.126284] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.126641] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] self.force_reraise() [ 656.126641] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.126641] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] raise self.value [ 656.126641] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 656.126641] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] updated_port = self._update_port( [ 656.126641] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.126641] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] _ensure_no_port_binding_failure(port) [ 656.126641] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.126641] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] raise exception.PortBindingFailed(port_id=port['id']) [ 656.126641] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] nova.exception.PortBindingFailed: Binding failed for port 3041ccf6-e87e-4e4e-8332-1a04de5361ce, please check neutron logs for more information. [ 656.126641] env[61987]: ERROR nova.compute.manager [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] [ 656.126896] env[61987]: DEBUG nova.compute.utils [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Binding failed for port 3041ccf6-e87e-4e4e-8332-1a04de5361ce, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 656.127503] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.438s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.130328] env[61987]: DEBUG nova.compute.manager [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Build of instance ab2edc70-b4c6-4db3-b0d6-53148764efdd was re-scheduled: Binding failed for port 3041ccf6-e87e-4e4e-8332-1a04de5361ce, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 656.131025] env[61987]: DEBUG nova.compute.manager [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 656.131025] env[61987]: DEBUG oslo_concurrency.lockutils [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Acquiring lock "refresh_cache-ab2edc70-b4c6-4db3-b0d6-53148764efdd" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.131146] env[61987]: DEBUG oslo_concurrency.lockutils [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Acquired lock "refresh_cache-ab2edc70-b4c6-4db3-b0d6-53148764efdd" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.131300] env[61987]: DEBUG nova.network.neutron [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 656.301304] env[61987]: INFO nova.compute.manager [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] [instance: 967bc997-543f-4b35-968b-c530b82d9db0] Took 1.02 seconds to deallocate network for instance. [ 656.464036] env[61987]: DEBUG nova.compute.manager [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 656.476392] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e6d5f05-e6f2-4b4b-a5dd-1086eb543fd5 tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Lock "6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.766s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.652682] env[61987]: DEBUG nova.network.neutron [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 656.742243] env[61987]: DEBUG nova.network.neutron [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.779215] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Acquiring lock "bb7243dc-a950-4d3e-9b17-944e57cdd56f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.779587] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Lock "bb7243dc-a950-4d3e-9b17-944e57cdd56f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.981515] env[61987]: DEBUG nova.compute.manager [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 656.985097] env[61987]: DEBUG oslo_concurrency.lockutils [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.016786] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5912a417-d2d2-4ef2-ad13-d8322a27bc3b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.024794] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78f73931-0849-4c1d-870d-48dc2066df71 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.054429] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66fb3b90-e564-4b7d-87f7-6817440ff6d3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.061066] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ec8d784-3c98-4b46-a605-f86fa5cec765 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.074250] env[61987]: DEBUG nova.compute.provider_tree [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 657.245975] env[61987]: DEBUG oslo_concurrency.lockutils [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Releasing lock "refresh_cache-ab2edc70-b4c6-4db3-b0d6-53148764efdd" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.246269] env[61987]: DEBUG nova.compute.manager [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 657.246521] env[61987]: DEBUG nova.compute.manager [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 657.246717] env[61987]: DEBUG nova.network.neutron [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 657.260881] env[61987]: DEBUG nova.network.neutron [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 657.316722] env[61987]: DEBUG oslo_concurrency.lockutils [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Acquiring lock "6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.316973] env[61987]: DEBUG oslo_concurrency.lockutils [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Lock "6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.317201] env[61987]: DEBUG oslo_concurrency.lockutils [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Acquiring lock "6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.317385] env[61987]: DEBUG oslo_concurrency.lockutils [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Lock "6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.317554] env[61987]: DEBUG oslo_concurrency.lockutils [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Lock "6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.319786] env[61987]: INFO nova.compute.manager [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Terminating instance [ 657.339286] env[61987]: INFO nova.scheduler.client.report [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Deleted allocations for instance 967bc997-543f-4b35-968b-c530b82d9db0 [ 657.506276] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.577506] env[61987]: DEBUG nova.scheduler.client.report [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 657.764220] env[61987]: DEBUG nova.network.neutron [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.824042] env[61987]: DEBUG nova.compute.manager [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 657.824042] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 657.824993] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f8e1acc-b14d-419e-873b-dca5fc46a8c5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.832694] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 657.833058] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d2d70b2-2109-488b-82b0-348630916e73 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.838630] env[61987]: DEBUG oslo_vmware.api [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Waiting for the task: (returnval){ [ 657.838630] env[61987]: value = "task-1061276" [ 657.838630] env[61987]: _type = "Task" [ 657.838630] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.847046] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc2628d5-2b98-448d-80bc-50ee318153b6 tempest-DeleteServersAdminTestJSON-1532307239 tempest-DeleteServersAdminTestJSON-1532307239-project-member] Lock "967bc997-543f-4b35-968b-c530b82d9db0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.726s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.847260] env[61987]: DEBUG oslo_vmware.api [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Task: {'id': task-1061276, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.082887] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.955s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.083604] env[61987]: ERROR nova.compute.manager [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 264cbccf-a94c-47d7-a785-9b2b8bd0fa58, please check neutron logs for more information. [ 658.083604] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Traceback (most recent call last): [ 658.083604] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 658.083604] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] self.driver.spawn(context, instance, image_meta, [ 658.083604] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 658.083604] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 658.083604] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 658.083604] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] vm_ref = self.build_virtual_machine(instance, [ 658.083604] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 658.083604] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] vif_infos = vmwarevif.get_vif_info(self._session, [ 658.083604] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 658.083909] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] for vif in network_info: [ 658.083909] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 658.083909] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] return self._sync_wrapper(fn, *args, **kwargs) [ 658.083909] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 658.083909] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] self.wait() [ 658.083909] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 658.083909] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] self[:] = self._gt.wait() [ 658.083909] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 658.083909] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] return self._exit_event.wait() [ 658.083909] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 658.083909] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] result = hub.switch() [ 658.083909] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 658.083909] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] return self.greenlet.switch() [ 658.084285] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.084285] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] result = function(*args, **kwargs) [ 658.084285] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 658.084285] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] return func(*args, **kwargs) [ 658.084285] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 658.084285] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] raise e [ 658.084285] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 658.084285] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] nwinfo = self.network_api.allocate_for_instance( [ 658.084285] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 658.084285] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] created_port_ids = self._update_ports_for_instance( [ 658.084285] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 658.084285] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] with excutils.save_and_reraise_exception(): [ 658.084285] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.084624] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] self.force_reraise() [ 658.084624] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.084624] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] raise self.value [ 658.084624] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 658.084624] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] updated_port = self._update_port( [ 658.084624] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.084624] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] _ensure_no_port_binding_failure(port) [ 658.084624] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.084624] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] raise exception.PortBindingFailed(port_id=port['id']) [ 658.084624] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] nova.exception.PortBindingFailed: Binding failed for port 264cbccf-a94c-47d7-a785-9b2b8bd0fa58, please check neutron logs for more information. [ 658.084624] env[61987]: ERROR nova.compute.manager [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] [ 658.084888] env[61987]: DEBUG nova.compute.utils [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Binding failed for port 264cbccf-a94c-47d7-a785-9b2b8bd0fa58, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 658.085590] env[61987]: DEBUG oslo_concurrency.lockutils [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.830s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.090432] env[61987]: DEBUG nova.compute.manager [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Build of instance 388b62aa-6647-4aca-acb2-8d7d7d424c8c was re-scheduled: Binding failed for port 264cbccf-a94c-47d7-a785-9b2b8bd0fa58, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 658.090432] env[61987]: DEBUG nova.compute.manager [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 658.090635] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Acquiring lock "refresh_cache-388b62aa-6647-4aca-acb2-8d7d7d424c8c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.090635] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Acquired lock "refresh_cache-388b62aa-6647-4aca-acb2-8d7d7d424c8c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.090772] env[61987]: DEBUG nova.network.neutron [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 658.267792] env[61987]: INFO nova.compute.manager [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] [instance: ab2edc70-b4c6-4db3-b0d6-53148764efdd] Took 1.02 seconds to deallocate network for instance. [ 658.348784] env[61987]: DEBUG oslo_vmware.api [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Task: {'id': task-1061276, 'name': PowerOffVM_Task, 'duration_secs': 0.203973} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.349440] env[61987]: DEBUG nova.compute.manager [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 658.351558] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 658.351736] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 658.352173] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c0c6be26-0f3f-42c8-9043-04820fa8eb89 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.416113] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 658.416113] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 658.416113] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Deleting the datastore file [datastore2] 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 658.416327] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-203acbe7-ddcb-42c6-b749-7ef04646ac99 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.422928] env[61987]: DEBUG oslo_vmware.api [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Waiting for the task: (returnval){ [ 658.422928] env[61987]: value = "task-1061278" [ 658.422928] env[61987]: _type = "Task" [ 658.422928] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.430907] env[61987]: DEBUG oslo_vmware.api [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Task: {'id': task-1061278, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.618495] env[61987]: DEBUG nova.network.neutron [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 658.742231] env[61987]: DEBUG nova.network.neutron [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.865854] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.932028] env[61987]: DEBUG oslo_vmware.api [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Task: {'id': task-1061278, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191823} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.932292] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 658.932491] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 658.932670] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 658.932842] env[61987]: INFO nova.compute.manager [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Took 1.11 seconds to destroy the instance on the hypervisor. [ 658.933092] env[61987]: DEBUG oslo.service.loopingcall [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 658.933305] env[61987]: DEBUG nova.compute.manager [-] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 658.933404] env[61987]: DEBUG nova.network.neutron [-] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 658.957463] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50552bdc-b975-4192-a972-4a626057936e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.964936] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-894f566c-5d2a-46df-861c-bb8000f1e9ac {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.997378] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4189d0e3-8b05-425f-983c-1fd04bf65500 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.004822] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f96ff58e-f5c7-4be5-a65b-8c898dfef4c4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.017762] env[61987]: DEBUG nova.compute.provider_tree [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.248638] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Releasing lock "refresh_cache-388b62aa-6647-4aca-acb2-8d7d7d424c8c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.248975] env[61987]: DEBUG nova.compute.manager [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 659.249229] env[61987]: DEBUG nova.compute.manager [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 659.249432] env[61987]: DEBUG nova.network.neutron [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 659.271615] env[61987]: DEBUG nova.network.neutron [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 659.312544] env[61987]: INFO nova.scheduler.client.report [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Deleted allocations for instance ab2edc70-b4c6-4db3-b0d6-53148764efdd [ 659.360886] env[61987]: DEBUG nova.compute.manager [req-755310cf-9905-47d6-956a-59e41be07ff6 req-73611841-b201-48e9-ad70-5daa2f1edc77 service nova] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Received event network-vif-deleted-90e59ebc-8840-4623-9c9b-2730ddf7f708 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 659.361104] env[61987]: INFO nova.compute.manager [req-755310cf-9905-47d6-956a-59e41be07ff6 req-73611841-b201-48e9-ad70-5daa2f1edc77 service nova] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Neutron deleted interface 90e59ebc-8840-4623-9c9b-2730ddf7f708; detaching it from the instance and deleting it from the info cache [ 659.361303] env[61987]: DEBUG nova.network.neutron [req-755310cf-9905-47d6-956a-59e41be07ff6 req-73611841-b201-48e9-ad70-5daa2f1edc77 service nova] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.521403] env[61987]: DEBUG nova.scheduler.client.report [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 659.775665] env[61987]: DEBUG nova.network.neutron [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.805723] env[61987]: DEBUG nova.network.neutron [-] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.823752] env[61987]: DEBUG oslo_concurrency.lockutils [None req-053fe0a0-08d7-4974-b4ca-de104c7ed7fb tempest-ServerPasswordTestJSON-1830787527 tempest-ServerPasswordTestJSON-1830787527-project-member] Lock "ab2edc70-b4c6-4db3-b0d6-53148764efdd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.686s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.864521] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d4febaa4-8ac3-4f09-aeac-6190cda872fe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.879194] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc676fac-3430-4c1b-b849-f154091a1656 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.910055] env[61987]: DEBUG nova.compute.manager [req-755310cf-9905-47d6-956a-59e41be07ff6 req-73611841-b201-48e9-ad70-5daa2f1edc77 service nova] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Detach interface failed, port_id=90e59ebc-8840-4623-9c9b-2730ddf7f708, reason: Instance 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 660.026802] env[61987]: DEBUG oslo_concurrency.lockutils [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.941s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.027455] env[61987]: ERROR nova.compute.manager [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 41fd9385-de81-4bd6-8feb-55408aca0c0a, please check neutron logs for more information. [ 660.027455] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Traceback (most recent call last): [ 660.027455] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 660.027455] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] self.driver.spawn(context, instance, image_meta, [ 660.027455] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 660.027455] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.027455] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.027455] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] vm_ref = self.build_virtual_machine(instance, [ 660.027455] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.027455] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.027455] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.027802] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] for vif in network_info: [ 660.027802] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 660.027802] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] return self._sync_wrapper(fn, *args, **kwargs) [ 660.027802] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 660.027802] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] self.wait() [ 660.027802] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 660.027802] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] self[:] = self._gt.wait() [ 660.027802] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.027802] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] return self._exit_event.wait() [ 660.027802] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 660.027802] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] current.throw(*self._exc) [ 660.027802] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.027802] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] result = function(*args, **kwargs) [ 660.028127] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 660.028127] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] return func(*args, **kwargs) [ 660.028127] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 660.028127] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] raise e [ 660.028127] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 660.028127] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] nwinfo = self.network_api.allocate_for_instance( [ 660.028127] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 660.028127] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] created_port_ids = self._update_ports_for_instance( [ 660.028127] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 660.028127] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] with excutils.save_and_reraise_exception(): [ 660.028127] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.028127] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] self.force_reraise() [ 660.028127] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.028426] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] raise self.value [ 660.028426] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 660.028426] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] updated_port = self._update_port( [ 660.028426] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.028426] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] _ensure_no_port_binding_failure(port) [ 660.028426] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.028426] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] raise exception.PortBindingFailed(port_id=port['id']) [ 660.028426] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] nova.exception.PortBindingFailed: Binding failed for port 41fd9385-de81-4bd6-8feb-55408aca0c0a, please check neutron logs for more information. [ 660.028426] env[61987]: ERROR nova.compute.manager [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] [ 660.028426] env[61987]: DEBUG nova.compute.utils [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Binding failed for port 41fd9385-de81-4bd6-8feb-55408aca0c0a, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 660.029963] env[61987]: DEBUG nova.compute.manager [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Build of instance 96bbc732-cb58-4de2-80d3-c527228adaea was re-scheduled: Binding failed for port 41fd9385-de81-4bd6-8feb-55408aca0c0a, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 660.030388] env[61987]: DEBUG nova.compute.manager [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 660.030619] env[61987]: DEBUG oslo_concurrency.lockutils [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Acquiring lock "refresh_cache-96bbc732-cb58-4de2-80d3-c527228adaea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.030769] env[61987]: DEBUG oslo_concurrency.lockutils [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Acquired lock "refresh_cache-96bbc732-cb58-4de2-80d3-c527228adaea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.030925] env[61987]: DEBUG nova.network.neutron [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 660.032375] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.318s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.278167] env[61987]: INFO nova.compute.manager [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] [instance: 388b62aa-6647-4aca-acb2-8d7d7d424c8c] Took 1.03 seconds to deallocate network for instance. [ 660.307923] env[61987]: INFO nova.compute.manager [-] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Took 1.37 seconds to deallocate network for instance. [ 660.327536] env[61987]: DEBUG nova.compute.manager [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 660.566574] env[61987]: DEBUG nova.network.neutron [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 660.795186] env[61987]: DEBUG nova.network.neutron [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.823587] env[61987]: DEBUG oslo_concurrency.lockutils [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.849320] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.937674] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec3e68d9-2220-4382-bd59-30e593b805d4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.945872] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12c13484-0c3f-4870-bff1-25688f2fe001 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.977256] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea783e62-6792-451e-8c46-533da5053923 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.984603] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1024f7ca-5755-4a57-9b00-af66588977dc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.999661] env[61987]: DEBUG nova.compute.provider_tree [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.297813] env[61987]: DEBUG oslo_concurrency.lockutils [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Releasing lock "refresh_cache-96bbc732-cb58-4de2-80d3-c527228adaea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.298068] env[61987]: DEBUG nova.compute.manager [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 661.298243] env[61987]: DEBUG nova.compute.manager [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 661.298401] env[61987]: DEBUG nova.network.neutron [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 661.309656] env[61987]: INFO nova.scheduler.client.report [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Deleted allocations for instance 388b62aa-6647-4aca-acb2-8d7d7d424c8c [ 661.377021] env[61987]: DEBUG nova.network.neutron [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 661.502803] env[61987]: DEBUG nova.scheduler.client.report [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 661.824699] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2244e3a3-4d6f-4e89-9f33-b38d600e6fcd tempest-ServerAddressesNegativeTestJSON-693741318 tempest-ServerAddressesNegativeTestJSON-693741318-project-member] Lock "388b62aa-6647-4aca-acb2-8d7d7d424c8c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.999s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.881197] env[61987]: DEBUG nova.network.neutron [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.009543] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.976s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.009543] env[61987]: ERROR nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 59570761-d84d-4350-9f6c-17502c6e58ff, please check neutron logs for more information. [ 662.009543] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Traceback (most recent call last): [ 662.009543] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 662.009543] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] self.driver.spawn(context, instance, image_meta, [ 662.009543] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 662.009543] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] self._vmops.spawn(context, instance, image_meta, injected_files, [ 662.009543] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 662.009543] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] vm_ref = self.build_virtual_machine(instance, [ 662.010110] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 662.010110] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] vif_infos = vmwarevif.get_vif_info(self._session, [ 662.010110] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 662.010110] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] for vif in network_info: [ 662.010110] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 662.010110] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] return self._sync_wrapper(fn, *args, **kwargs) [ 662.010110] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 662.010110] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] self.wait() [ 662.010110] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 662.010110] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] self[:] = self._gt.wait() [ 662.010110] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 662.010110] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] return self._exit_event.wait() [ 662.010110] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 662.010453] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] result = hub.switch() [ 662.010453] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 662.010453] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] return self.greenlet.switch() [ 662.010453] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.010453] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] result = function(*args, **kwargs) [ 662.010453] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 662.010453] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] return func(*args, **kwargs) [ 662.010453] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 662.010453] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] raise e [ 662.010453] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 662.010453] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] nwinfo = self.network_api.allocate_for_instance( [ 662.010453] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 662.010453] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] created_port_ids = self._update_ports_for_instance( [ 662.010783] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 662.010783] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] with excutils.save_and_reraise_exception(): [ 662.010783] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.010783] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] self.force_reraise() [ 662.010783] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.010783] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] raise self.value [ 662.010783] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 662.010783] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] updated_port = self._update_port( [ 662.010783] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.010783] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] _ensure_no_port_binding_failure(port) [ 662.010783] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.010783] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] raise exception.PortBindingFailed(port_id=port['id']) [ 662.011213] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] nova.exception.PortBindingFailed: Binding failed for port 59570761-d84d-4350-9f6c-17502c6e58ff, please check neutron logs for more information. [ 662.011213] env[61987]: ERROR nova.compute.manager [instance: 699c853a-a88f-4890-b712-00f00a6c9838] [ 662.011213] env[61987]: DEBUG nova.compute.utils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Binding failed for port 59570761-d84d-4350-9f6c-17502c6e58ff, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 662.015020] env[61987]: DEBUG nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Build of instance 699c853a-a88f-4890-b712-00f00a6c9838 was re-scheduled: Binding failed for port 59570761-d84d-4350-9f6c-17502c6e58ff, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 662.015020] env[61987]: DEBUG nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 662.015020] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "refresh_cache-699c853a-a88f-4890-b712-00f00a6c9838" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.015020] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquired lock "refresh_cache-699c853a-a88f-4890-b712-00f00a6c9838" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.015538] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 662.015538] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.041s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.333024] env[61987]: DEBUG nova.compute.manager [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 662.386107] env[61987]: INFO nova.compute.manager [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] [instance: 96bbc732-cb58-4de2-80d3-c527228adaea] Took 1.09 seconds to deallocate network for instance. [ 662.552931] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.754098] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.860375] env[61987]: DEBUG oslo_concurrency.lockutils [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.989710] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b811d9-ae5d-4bcd-aabd-059cf03569c0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.998576] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d9bf92-547e-4234-96cd-efb420ec70b1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.036231] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0189fa13-381c-462d-9671-4aa778b69682 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.045047] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fe8b306-c093-4469-840e-3cefba1edcf5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.061021] env[61987]: DEBUG nova.compute.provider_tree [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.260226] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Releasing lock "refresh_cache-699c853a-a88f-4890-b712-00f00a6c9838" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.260496] env[61987]: DEBUG nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 663.260680] env[61987]: DEBUG nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 663.260850] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 663.276887] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 663.435434] env[61987]: INFO nova.scheduler.client.report [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Deleted allocations for instance 96bbc732-cb58-4de2-80d3-c527228adaea [ 663.566990] env[61987]: DEBUG nova.scheduler.client.report [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 663.783121] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.947635] env[61987]: DEBUG oslo_concurrency.lockutils [None req-02524f38-0eb0-415e-9633-ac3120ea3df8 tempest-ServersTestFqdnHostnames-1268586742 tempest-ServersTestFqdnHostnames-1268586742-project-member] Lock "96bbc732-cb58-4de2-80d3-c527228adaea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 99.526s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.072045] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.056s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.072893] env[61987]: ERROR nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 26ebdb7d-365e-4aa3-a0ae-35812dd80bbb, please check neutron logs for more information. [ 664.072893] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Traceback (most recent call last): [ 664.072893] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 664.072893] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] self.driver.spawn(context, instance, image_meta, [ 664.072893] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 664.072893] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 664.072893] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 664.072893] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] vm_ref = self.build_virtual_machine(instance, [ 664.072893] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 664.072893] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] vif_infos = vmwarevif.get_vif_info(self._session, [ 664.072893] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 664.073383] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] for vif in network_info: [ 664.073383] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 664.073383] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] return self._sync_wrapper(fn, *args, **kwargs) [ 664.073383] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 664.073383] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] self.wait() [ 664.073383] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 664.073383] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] self[:] = self._gt.wait() [ 664.073383] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 664.073383] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] return self._exit_event.wait() [ 664.073383] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 664.073383] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] current.throw(*self._exc) [ 664.073383] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.073383] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] result = function(*args, **kwargs) [ 664.074051] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 664.074051] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] return func(*args, **kwargs) [ 664.074051] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 664.074051] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] raise e [ 664.074051] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 664.074051] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] nwinfo = self.network_api.allocate_for_instance( [ 664.074051] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 664.074051] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] created_port_ids = self._update_ports_for_instance( [ 664.074051] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 664.074051] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] with excutils.save_and_reraise_exception(): [ 664.074051] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.074051] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] self.force_reraise() [ 664.074051] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.074720] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] raise self.value [ 664.074720] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 664.074720] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] updated_port = self._update_port( [ 664.074720] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.074720] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] _ensure_no_port_binding_failure(port) [ 664.074720] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.074720] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] raise exception.PortBindingFailed(port_id=port['id']) [ 664.074720] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] nova.exception.PortBindingFailed: Binding failed for port 26ebdb7d-365e-4aa3-a0ae-35812dd80bbb, please check neutron logs for more information. [ 664.074720] env[61987]: ERROR nova.compute.manager [instance: 65ddff8e-971c-4485-b658-4611aea1176c] [ 664.074720] env[61987]: DEBUG nova.compute.utils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Binding failed for port 26ebdb7d-365e-4aa3-a0ae-35812dd80bbb, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 664.075040] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.729s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.080020] env[61987]: DEBUG nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Build of instance 65ddff8e-971c-4485-b658-4611aea1176c was re-scheduled: Binding failed for port 26ebdb7d-365e-4aa3-a0ae-35812dd80bbb, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 664.080020] env[61987]: DEBUG nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 664.080020] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "refresh_cache-65ddff8e-971c-4485-b658-4611aea1176c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.080020] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquired lock "refresh_cache-65ddff8e-971c-4485-b658-4611aea1176c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.080320] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 664.286410] env[61987]: INFO nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 699c853a-a88f-4890-b712-00f00a6c9838] Took 1.03 seconds to deallocate network for instance. [ 664.451491] env[61987]: DEBUG nova.compute.manager [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 664.605139] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 664.708755] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.977175] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.215671] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Releasing lock "refresh_cache-65ddff8e-971c-4485-b658-4611aea1176c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.215671] env[61987]: DEBUG nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 665.215671] env[61987]: DEBUG nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 665.215671] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 665.263209] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 665.322192] env[61987]: INFO nova.scheduler.client.report [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Deleted allocations for instance 699c853a-a88f-4890-b712-00f00a6c9838 [ 665.620656] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 699c853a-a88f-4890-b712-00f00a6c9838 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 665.764998] env[61987]: DEBUG nova.network.neutron [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.832736] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "699c853a-a88f-4890-b712-00f00a6c9838" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.287s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.124991] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 65ddff8e-971c-4485-b658-4611aea1176c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 666.125217] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 666.125417] env[61987]: WARNING nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 666.267899] env[61987]: INFO nova.compute.manager [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 65ddff8e-971c-4485-b658-4611aea1176c] Took 1.05 seconds to deallocate network for instance. [ 666.336911] env[61987]: DEBUG nova.compute.manager [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 666.630285] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 89a97f81-bcd9-49e5-bbd3-9140f6638a20 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 666.866031] env[61987]: DEBUG oslo_concurrency.lockutils [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.132348] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 630e0959-6d7f-401d-b84c-1c098f54bf60 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 667.301396] env[61987]: INFO nova.scheduler.client.report [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Deleted allocations for instance 65ddff8e-971c-4485-b658-4611aea1176c [ 667.636753] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 466a5f5a-f48a-4921-8bcb-44a1840c8141 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 667.811171] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7423af7d-f6ba-4414-af6d-a31e6c2841fd tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "65ddff8e-971c-4485-b658-4611aea1176c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.237s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.139981] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance ea6c2120-a69a-42c7-bc92-3e74539c008d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 668.315593] env[61987]: DEBUG nova.compute.manager [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 668.647832] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 02fe490b-75c8-406b-98e0-aa200f249673 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 668.846375] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.152911] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 0b4b3bdb-7393-47ab-abc1-44c47496cb3a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 669.475489] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "f937a5ec-a1d7-41d8-b998-fa18f545b304" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.475489] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "f937a5ec-a1d7-41d8-b998-fa18f545b304" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.658516] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 51a36102-795e-47b7-b96a-857e54dc703e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 670.162665] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance fa943137-caf1-4b41-b644-c9146f72f956 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 670.168273] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "d566ea53-958b-4a35-989c-771180d95bb4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.168652] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "d566ea53-958b-4a35-989c-771180d95bb4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.206444] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "8e68c5a2-6587-45b6-9a76-96b129a0a665" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.206933] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "8e68c5a2-6587-45b6-9a76-96b129a0a665" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.665345] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 671.169925] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 671.671713] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 672.175543] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 672.677464] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 93174a10-f8b1-4789-ab3c-dda07bdddaa8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 673.180617] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance d55395cc-0b92-4e99-9b80-913c19f6c7f4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 673.685033] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 4edd0159-6db1-41bd-a616-31a34e984059 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 674.187467] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 325183b9-e35b-4f9c-a271-d0fdb05288bb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 674.691030] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance d038391c-2b32-455c-85d5-68221e7bd065 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 675.167962] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Acquiring lock "1c63bf62-3f70-42b9-b6d7-41336d20f6c8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.168206] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Lock "1c63bf62-3f70-42b9-b6d7-41336d20f6c8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.196053] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance ad40937f-d5a9-4708-8b81-06087d38d765 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 675.699025] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 1afcffda-0bd2-4cbe-8cfb-12a91bb50975 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 676.202223] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 95641d0b-970c-4b94-8568-5c46d0808345 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 676.705820] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 677.208729] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance bb7243dc-a950-4d3e-9b17-944e57cdd56f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 677.208916] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 677.209071] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 677.576240] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4c562df-30c0-47c7-a5a3-cb546235faec {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.584044] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d9bb2d-852c-471b-93e5-484174a34f12 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.613027] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba38dec2-9225-4ed7-b40c-01aa17c191bf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.620198] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fe8fd59-9797-4e84-9e99-ea89de54a0a2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.632885] env[61987]: DEBUG nova.compute.provider_tree [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.136134] env[61987]: DEBUG nova.scheduler.client.report [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 678.640783] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61987) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 678.641076] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 14.566s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.641372] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.333s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.428180] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed12d24-ad76-45fc-a739-d5668b09643f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.435782] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8ce285e-5436-4ad7-a534-3249a8afc92c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.465560] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba20d69-3615-45b9-bb5f-d07b93cdcf99 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.473084] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95222f42-8516-4094-aa6a-dd982867e935 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.485910] env[61987]: DEBUG nova.compute.provider_tree [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.989060] env[61987]: DEBUG nova.scheduler.client.report [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 680.494954] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.853s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.495681] env[61987]: ERROR nova.compute.manager [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bf22278e-072b-41be-bcd4-ec58e771be7d, please check neutron logs for more information. [ 680.495681] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Traceback (most recent call last): [ 680.495681] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 680.495681] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] self.driver.spawn(context, instance, image_meta, [ 680.495681] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 680.495681] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.495681] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.495681] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] vm_ref = self.build_virtual_machine(instance, [ 680.495681] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.495681] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.495681] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.495971] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] for vif in network_info: [ 680.495971] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.495971] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] return self._sync_wrapper(fn, *args, **kwargs) [ 680.495971] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.495971] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] self.wait() [ 680.495971] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.495971] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] self[:] = self._gt.wait() [ 680.495971] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.495971] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] return self._exit_event.wait() [ 680.495971] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 680.495971] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] result = hub.switch() [ 680.495971] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 680.495971] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] return self.greenlet.switch() [ 680.496397] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.496397] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] result = function(*args, **kwargs) [ 680.496397] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.496397] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] return func(*args, **kwargs) [ 680.496397] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 680.496397] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] raise e [ 680.496397] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 680.496397] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] nwinfo = self.network_api.allocate_for_instance( [ 680.496397] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.496397] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] created_port_ids = self._update_ports_for_instance( [ 680.496397] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.496397] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] with excutils.save_and_reraise_exception(): [ 680.496397] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.496715] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] self.force_reraise() [ 680.496715] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.496715] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] raise self.value [ 680.496715] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.496715] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] updated_port = self._update_port( [ 680.496715] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.496715] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] _ensure_no_port_binding_failure(port) [ 680.496715] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.496715] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] raise exception.PortBindingFailed(port_id=port['id']) [ 680.496715] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] nova.exception.PortBindingFailed: Binding failed for port bf22278e-072b-41be-bcd4-ec58e771be7d, please check neutron logs for more information. [ 680.496715] env[61987]: ERROR nova.compute.manager [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] [ 680.496999] env[61987]: DEBUG nova.compute.utils [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Binding failed for port bf22278e-072b-41be-bcd4-ec58e771be7d, please check neutron logs for more information. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 680.497741] env[61987]: DEBUG oslo_concurrency.lockutils [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.451s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.499205] env[61987]: INFO nova.compute.claims [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 680.501785] env[61987]: DEBUG nova.compute.manager [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Build of instance f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0 was re-scheduled: Binding failed for port bf22278e-072b-41be-bcd4-ec58e771be7d, please check neutron logs for more information. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 680.502221] env[61987]: DEBUG nova.compute.manager [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Unplugging VIFs for instance {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 680.502443] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Acquiring lock "refresh_cache-f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.502592] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Acquired lock "refresh_cache-f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.502749] env[61987]: DEBUG nova.network.neutron [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 681.024919] env[61987]: DEBUG nova.network.neutron [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.091525] env[61987]: DEBUG nova.network.neutron [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.594356] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Releasing lock "refresh_cache-f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.594659] env[61987]: DEBUG nova.compute.manager [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61987) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 681.594827] env[61987]: DEBUG nova.compute.manager [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 681.594998] env[61987]: DEBUG nova.network.neutron [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 681.614172] env[61987]: DEBUG nova.network.neutron [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.818152] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a78cab9-9adf-478b-84e2-d4fd2f4572b2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.825183] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa3006b6-00de-487d-a543-aee81bd09d67 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.853866] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04b9b642-3bfb-4761-a588-d913a461610b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.860747] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56c6f3a5-88d2-45e7-a216-2286ea78de6b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.873638] env[61987]: DEBUG nova.compute.provider_tree [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 682.117132] env[61987]: DEBUG nova.network.neutron [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.377225] env[61987]: DEBUG nova.scheduler.client.report [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 682.620196] env[61987]: INFO nova.compute.manager [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0] Took 1.02 seconds to deallocate network for instance. [ 682.883020] env[61987]: DEBUG oslo_concurrency.lockutils [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.385s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.883160] env[61987]: DEBUG nova.compute.manager [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 682.886024] env[61987]: DEBUG oslo_concurrency.lockutils [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.901s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.887231] env[61987]: INFO nova.compute.claims [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 683.391705] env[61987]: DEBUG nova.compute.utils [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 683.395200] env[61987]: DEBUG nova.compute.manager [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 683.395384] env[61987]: DEBUG nova.network.neutron [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 683.447677] env[61987]: DEBUG nova.policy [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cc4d69a44feb49c18c22505a8f4e84c9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8c17c124c6c74011894c896a918534d1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 683.651816] env[61987]: INFO nova.scheduler.client.report [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Deleted allocations for instance f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0 [ 683.757750] env[61987]: DEBUG nova.network.neutron [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Successfully created port: cf9c4ef3-c7ef-4194-bd5b-72f4bafee5fa {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 683.899030] env[61987]: DEBUG nova.compute.manager [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 684.160486] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae52f66e-7a76-4406-8850-f623f416c218 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Lock "f6ebfa11-c9b6-46dd-8ea4-1ee0d32484e0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.014s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.266755] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8dcf47b-6bac-476f-8c18-3d333848f32f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.274444] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d069df6-76d9-479a-965e-f64757290c2b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.305024] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded7bcc9-6de5-498a-b656-5b6c155741dc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.312215] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-148b9f47-a3fe-4e35-8507-13f795725c07 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.325072] env[61987]: DEBUG nova.compute.provider_tree [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.666689] env[61987]: DEBUG nova.compute.manager [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 684.827648] env[61987]: DEBUG nova.scheduler.client.report [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 684.909584] env[61987]: DEBUG nova.compute.manager [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 684.933677] env[61987]: DEBUG nova.virt.hardware [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 684.933989] env[61987]: DEBUG nova.virt.hardware [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 684.934183] env[61987]: DEBUG nova.virt.hardware [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 684.934499] env[61987]: DEBUG nova.virt.hardware [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 684.934689] env[61987]: DEBUG nova.virt.hardware [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 684.934846] env[61987]: DEBUG nova.virt.hardware [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 684.935068] env[61987]: DEBUG nova.virt.hardware [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 684.935235] env[61987]: DEBUG nova.virt.hardware [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 684.935404] env[61987]: DEBUG nova.virt.hardware [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 684.935572] env[61987]: DEBUG nova.virt.hardware [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 684.935748] env[61987]: DEBUG nova.virt.hardware [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 684.936708] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-252c2991-58c5-4f1b-bd4d-c0d62f6d473a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.944617] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2b54dca-46eb-46b5-87e2-50bdbbe0a7cd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.189144] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.270878] env[61987]: DEBUG nova.compute.manager [req-c91b8e27-40a6-4043-8fa9-81c985d02d24 req-af556801-603a-49d5-904c-5e57e2112261 service nova] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Received event network-vif-plugged-cf9c4ef3-c7ef-4194-bd5b-72f4bafee5fa {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 685.271102] env[61987]: DEBUG oslo_concurrency.lockutils [req-c91b8e27-40a6-4043-8fa9-81c985d02d24 req-af556801-603a-49d5-904c-5e57e2112261 service nova] Acquiring lock "89a97f81-bcd9-49e5-bbd3-9140f6638a20-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.271309] env[61987]: DEBUG oslo_concurrency.lockutils [req-c91b8e27-40a6-4043-8fa9-81c985d02d24 req-af556801-603a-49d5-904c-5e57e2112261 service nova] Lock "89a97f81-bcd9-49e5-bbd3-9140f6638a20-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.271504] env[61987]: DEBUG oslo_concurrency.lockutils [req-c91b8e27-40a6-4043-8fa9-81c985d02d24 req-af556801-603a-49d5-904c-5e57e2112261 service nova] Lock "89a97f81-bcd9-49e5-bbd3-9140f6638a20-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.271684] env[61987]: DEBUG nova.compute.manager [req-c91b8e27-40a6-4043-8fa9-81c985d02d24 req-af556801-603a-49d5-904c-5e57e2112261 service nova] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] No waiting events found dispatching network-vif-plugged-cf9c4ef3-c7ef-4194-bd5b-72f4bafee5fa {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 685.271846] env[61987]: WARNING nova.compute.manager [req-c91b8e27-40a6-4043-8fa9-81c985d02d24 req-af556801-603a-49d5-904c-5e57e2112261 service nova] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Received unexpected event network-vif-plugged-cf9c4ef3-c7ef-4194-bd5b-72f4bafee5fa for instance with vm_state building and task_state spawning. [ 685.334320] env[61987]: DEBUG oslo_concurrency.lockutils [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.334873] env[61987]: DEBUG nova.compute.manager [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 685.337541] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.831s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.339458] env[61987]: INFO nova.compute.claims [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 685.383295] env[61987]: DEBUG nova.network.neutron [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Successfully updated port: cf9c4ef3-c7ef-4194-bd5b-72f4bafee5fa {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 685.843831] env[61987]: DEBUG nova.compute.utils [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 685.847609] env[61987]: DEBUG nova.compute.manager [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 685.847698] env[61987]: DEBUG nova.network.neutron [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 685.889569] env[61987]: DEBUG oslo_concurrency.lockutils [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Acquiring lock "refresh_cache-89a97f81-bcd9-49e5-bbd3-9140f6638a20" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.889731] env[61987]: DEBUG oslo_concurrency.lockutils [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Acquired lock "refresh_cache-89a97f81-bcd9-49e5-bbd3-9140f6638a20" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.889895] env[61987]: DEBUG nova.network.neutron [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 685.898349] env[61987]: DEBUG nova.policy [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c33abbbe46e94874a4a5cc0cb4ee1cf0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '04f95c6e34db49c58e3827d7a72cab54', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 686.160603] env[61987]: DEBUG nova.network.neutron [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Successfully created port: 0c62bc7d-9ca1-42c9-b59b-53a18c465220 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 686.348191] env[61987]: DEBUG nova.compute.manager [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 686.624021] env[61987]: DEBUG nova.network.neutron [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 686.754958] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b73b05-0c72-4e8f-b2ac-21dcf20fbc9f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.769426] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f40ed36f-f864-41ab-9d55-0ac795a9dee2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.805336] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0569e447-37bc-4c26-8523-685c451cfe14 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.814041] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf105e62-1f38-4c50-a51d-d816d5869408 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.827270] env[61987]: DEBUG nova.compute.provider_tree [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.855425] env[61987]: DEBUG nova.network.neutron [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Updating instance_info_cache with network_info: [{"id": "cf9c4ef3-c7ef-4194-bd5b-72f4bafee5fa", "address": "fa:16:3e:b4:50:15", "network": {"id": "8a9770a4-fed3-4232-ac78-8a1fc2223229", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.26", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "705320a8b6a549d0b197ee1a35404d52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bafe8721-91d4-4127-b215-d9e8e27947dc", "external-id": "nsx-vlan-transportzone-680", "segmentation_id": 680, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf9c4ef3-c7", "ovs_interfaceid": "cf9c4ef3-c7ef-4194-bd5b-72f4bafee5fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.298374] env[61987]: DEBUG nova.compute.manager [req-06a3376a-6bd0-488d-8dde-39dd221ce96c req-d33b5218-ddff-4c98-8dff-7085064f4a9e service nova] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Received event network-changed-cf9c4ef3-c7ef-4194-bd5b-72f4bafee5fa {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 687.298562] env[61987]: DEBUG nova.compute.manager [req-06a3376a-6bd0-488d-8dde-39dd221ce96c req-d33b5218-ddff-4c98-8dff-7085064f4a9e service nova] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Refreshing instance network info cache due to event network-changed-cf9c4ef3-c7ef-4194-bd5b-72f4bafee5fa. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 687.298799] env[61987]: DEBUG oslo_concurrency.lockutils [req-06a3376a-6bd0-488d-8dde-39dd221ce96c req-d33b5218-ddff-4c98-8dff-7085064f4a9e service nova] Acquiring lock "refresh_cache-89a97f81-bcd9-49e5-bbd3-9140f6638a20" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.331388] env[61987]: DEBUG nova.scheduler.client.report [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 687.360212] env[61987]: DEBUG oslo_concurrency.lockutils [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Releasing lock "refresh_cache-89a97f81-bcd9-49e5-bbd3-9140f6638a20" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.360501] env[61987]: DEBUG nova.compute.manager [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Instance network_info: |[{"id": "cf9c4ef3-c7ef-4194-bd5b-72f4bafee5fa", "address": "fa:16:3e:b4:50:15", "network": {"id": "8a9770a4-fed3-4232-ac78-8a1fc2223229", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.26", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "705320a8b6a549d0b197ee1a35404d52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bafe8721-91d4-4127-b215-d9e8e27947dc", "external-id": "nsx-vlan-transportzone-680", "segmentation_id": 680, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf9c4ef3-c7", "ovs_interfaceid": "cf9c4ef3-c7ef-4194-bd5b-72f4bafee5fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 687.361771] env[61987]: DEBUG nova.compute.manager [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 687.364035] env[61987]: DEBUG oslo_concurrency.lockutils [req-06a3376a-6bd0-488d-8dde-39dd221ce96c req-d33b5218-ddff-4c98-8dff-7085064f4a9e service nova] Acquired lock "refresh_cache-89a97f81-bcd9-49e5-bbd3-9140f6638a20" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.364035] env[61987]: DEBUG nova.network.neutron [req-06a3376a-6bd0-488d-8dde-39dd221ce96c req-d33b5218-ddff-4c98-8dff-7085064f4a9e service nova] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Refreshing network info cache for port cf9c4ef3-c7ef-4194-bd5b-72f4bafee5fa {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 687.365036] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b4:50:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bafe8721-91d4-4127-b215-d9e8e27947dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cf9c4ef3-c7ef-4194-bd5b-72f4bafee5fa', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 687.372367] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Creating folder: Project (8c17c124c6c74011894c896a918534d1). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 687.375821] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2a3a61dc-234b-4c2e-afd4-91252d651b42 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.387983] env[61987]: DEBUG nova.virt.hardware [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 687.388275] env[61987]: DEBUG nova.virt.hardware [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 687.388464] env[61987]: DEBUG nova.virt.hardware [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 687.388683] env[61987]: DEBUG nova.virt.hardware [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 687.388914] env[61987]: DEBUG nova.virt.hardware [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 687.389071] env[61987]: DEBUG nova.virt.hardware [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 687.389380] env[61987]: DEBUG nova.virt.hardware [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 687.389534] env[61987]: DEBUG nova.virt.hardware [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 687.389699] env[61987]: DEBUG nova.virt.hardware [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 687.389921] env[61987]: DEBUG nova.virt.hardware [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 687.390200] env[61987]: DEBUG nova.virt.hardware [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 687.391233] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1ed03cb-1c4b-4e1f-b69d-797eecf9c7a9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.396542] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Created folder: Project (8c17c124c6c74011894c896a918534d1) in parent group-v234219. [ 687.396542] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Creating folder: Instances. Parent ref: group-v234233. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 687.397267] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f340e0ad-a8cc-4540-b4c6-975d7a762a77 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.401994] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d9db99e-7f3e-4299-a41e-d86612f88ff7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.406999] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Created folder: Instances in parent group-v234233. [ 687.407196] env[61987]: DEBUG oslo.service.loopingcall [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 687.409664] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 687.409904] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b6856df8-f6f9-442b-9907-702e16a81fa6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.438438] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 687.438438] env[61987]: value = "task-1061281" [ 687.438438] env[61987]: _type = "Task" [ 687.438438] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.446177] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061281, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.637341] env[61987]: DEBUG nova.network.neutron [req-06a3376a-6bd0-488d-8dde-39dd221ce96c req-d33b5218-ddff-4c98-8dff-7085064f4a9e service nova] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Updated VIF entry in instance network info cache for port cf9c4ef3-c7ef-4194-bd5b-72f4bafee5fa. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 687.637836] env[61987]: DEBUG nova.network.neutron [req-06a3376a-6bd0-488d-8dde-39dd221ce96c req-d33b5218-ddff-4c98-8dff-7085064f4a9e service nova] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Updating instance_info_cache with network_info: [{"id": "cf9c4ef3-c7ef-4194-bd5b-72f4bafee5fa", "address": "fa:16:3e:b4:50:15", "network": {"id": "8a9770a4-fed3-4232-ac78-8a1fc2223229", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.26", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "705320a8b6a549d0b197ee1a35404d52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bafe8721-91d4-4127-b215-d9e8e27947dc", "external-id": "nsx-vlan-transportzone-680", "segmentation_id": 680, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf9c4ef3-c7", "ovs_interfaceid": "cf9c4ef3-c7ef-4194-bd5b-72f4bafee5fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.835855] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.498s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.836496] env[61987]: DEBUG nova.compute.manager [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 687.839227] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.973s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.840612] env[61987]: INFO nova.compute.claims [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 687.949182] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061281, 'name': CreateVM_Task, 'duration_secs': 0.297018} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.949474] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 687.950186] env[61987]: DEBUG oslo_concurrency.lockutils [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.950421] env[61987]: DEBUG oslo_concurrency.lockutils [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.954019] env[61987]: DEBUG oslo_concurrency.lockutils [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 687.954019] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21aa03a5-9b91-4be6-936e-6e65a9d5bf72 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.955838] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Waiting for the task: (returnval){ [ 687.955838] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52914b65-af0f-10c9-6902-fb9429d996a8" [ 687.955838] env[61987]: _type = "Task" [ 687.955838] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.964035] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52914b65-af0f-10c9-6902-fb9429d996a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.060127] env[61987]: DEBUG nova.network.neutron [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Successfully updated port: 0c62bc7d-9ca1-42c9-b59b-53a18c465220 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 688.140420] env[61987]: DEBUG oslo_concurrency.lockutils [req-06a3376a-6bd0-488d-8dde-39dd221ce96c req-d33b5218-ddff-4c98-8dff-7085064f4a9e service nova] Releasing lock "refresh_cache-89a97f81-bcd9-49e5-bbd3-9140f6638a20" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.345031] env[61987]: DEBUG nova.compute.utils [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 688.346375] env[61987]: DEBUG nova.compute.manager [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 688.346548] env[61987]: DEBUG nova.network.neutron [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 688.385893] env[61987]: DEBUG nova.policy [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4d582141d7f4468d9f4f073f0090726f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd3adce5cde814bf0bb57b8e941dba79f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 688.466860] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52914b65-af0f-10c9-6902-fb9429d996a8, 'name': SearchDatastore_Task, 'duration_secs': 0.009435} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.467167] env[61987]: DEBUG oslo_concurrency.lockutils [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.467451] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 688.467945] env[61987]: DEBUG oslo_concurrency.lockutils [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.468164] env[61987]: DEBUG oslo_concurrency.lockutils [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.468365] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 688.468995] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4b9f8b2e-7f64-4626-a3c8-4c9c5aa4f0a0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.476971] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 688.477184] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 688.477891] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea1c37cd-7555-419f-823e-ae8704b9615b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.484330] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Waiting for the task: (returnval){ [ 688.484330] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5290a08a-4243-0cd7-7fdc-c88b6b7e9cf8" [ 688.484330] env[61987]: _type = "Task" [ 688.484330] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.492319] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5290a08a-4243-0cd7-7fdc-c88b6b7e9cf8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.562248] env[61987]: DEBUG oslo_concurrency.lockutils [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Acquiring lock "refresh_cache-630e0959-6d7f-401d-b84c-1c098f54bf60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.562388] env[61987]: DEBUG oslo_concurrency.lockutils [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Acquired lock "refresh_cache-630e0959-6d7f-401d-b84c-1c098f54bf60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.562542] env[61987]: DEBUG nova.network.neutron [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 688.727133] env[61987]: DEBUG nova.network.neutron [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Successfully created port: 39a5fbe3-c262-40f4-acdc-48209a7f617a {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 688.850379] env[61987]: DEBUG nova.compute.manager [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 688.997814] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5290a08a-4243-0cd7-7fdc-c88b6b7e9cf8, 'name': SearchDatastore_Task, 'duration_secs': 0.008437} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.999368] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03b26915-baab-4370-a7cc-44628de76b30 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.004147] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Waiting for the task: (returnval){ [ 689.004147] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5275d83c-b820-621a-9c19-69bf87084e93" [ 689.004147] env[61987]: _type = "Task" [ 689.004147] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.013855] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5275d83c-b820-621a-9c19-69bf87084e93, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.100383] env[61987]: DEBUG nova.network.neutron [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 689.191260] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5019783-0d3e-4e91-bc98-9fab96c2e20f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.199663] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41b452cb-3bb0-457a-9583-7f8e2fb2fa8c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.232013] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aaafb4e-a22e-4d90-b074-57936090b384 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.234974] env[61987]: DEBUG nova.network.neutron [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Updating instance_info_cache with network_info: [{"id": "0c62bc7d-9ca1-42c9-b59b-53a18c465220", "address": "fa:16:3e:71:6e:ef", "network": {"id": "c3010fca-d3b0-4624-b84b-b0920f0a0e05", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1462858397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04f95c6e34db49c58e3827d7a72cab54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c62bc7d-9c", "ovs_interfaceid": "0c62bc7d-9ca1-42c9-b59b-53a18c465220", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.240708] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b1dc370-1021-4f53-865d-d8af52f244ce {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.253740] env[61987]: DEBUG nova.compute.provider_tree [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.322978] env[61987]: DEBUG nova.compute.manager [req-3e5f6fa6-a6cd-4b9a-8c07-472dfcc32ce0 req-ca57d581-7b9d-40f7-bc75-cf93cad75a00 service nova] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Received event network-vif-plugged-0c62bc7d-9ca1-42c9-b59b-53a18c465220 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 689.323223] env[61987]: DEBUG oslo_concurrency.lockutils [req-3e5f6fa6-a6cd-4b9a-8c07-472dfcc32ce0 req-ca57d581-7b9d-40f7-bc75-cf93cad75a00 service nova] Acquiring lock "630e0959-6d7f-401d-b84c-1c098f54bf60-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.323476] env[61987]: DEBUG oslo_concurrency.lockutils [req-3e5f6fa6-a6cd-4b9a-8c07-472dfcc32ce0 req-ca57d581-7b9d-40f7-bc75-cf93cad75a00 service nova] Lock "630e0959-6d7f-401d-b84c-1c098f54bf60-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.323654] env[61987]: DEBUG oslo_concurrency.lockutils [req-3e5f6fa6-a6cd-4b9a-8c07-472dfcc32ce0 req-ca57d581-7b9d-40f7-bc75-cf93cad75a00 service nova] Lock "630e0959-6d7f-401d-b84c-1c098f54bf60-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.323823] env[61987]: DEBUG nova.compute.manager [req-3e5f6fa6-a6cd-4b9a-8c07-472dfcc32ce0 req-ca57d581-7b9d-40f7-bc75-cf93cad75a00 service nova] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] No waiting events found dispatching network-vif-plugged-0c62bc7d-9ca1-42c9-b59b-53a18c465220 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 689.323985] env[61987]: WARNING nova.compute.manager [req-3e5f6fa6-a6cd-4b9a-8c07-472dfcc32ce0 req-ca57d581-7b9d-40f7-bc75-cf93cad75a00 service nova] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Received unexpected event network-vif-plugged-0c62bc7d-9ca1-42c9-b59b-53a18c465220 for instance with vm_state building and task_state spawning. [ 689.324156] env[61987]: DEBUG nova.compute.manager [req-3e5f6fa6-a6cd-4b9a-8c07-472dfcc32ce0 req-ca57d581-7b9d-40f7-bc75-cf93cad75a00 service nova] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Received event network-changed-0c62bc7d-9ca1-42c9-b59b-53a18c465220 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 689.324308] env[61987]: DEBUG nova.compute.manager [req-3e5f6fa6-a6cd-4b9a-8c07-472dfcc32ce0 req-ca57d581-7b9d-40f7-bc75-cf93cad75a00 service nova] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Refreshing instance network info cache due to event network-changed-0c62bc7d-9ca1-42c9-b59b-53a18c465220. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 689.324501] env[61987]: DEBUG oslo_concurrency.lockutils [req-3e5f6fa6-a6cd-4b9a-8c07-472dfcc32ce0 req-ca57d581-7b9d-40f7-bc75-cf93cad75a00 service nova] Acquiring lock "refresh_cache-630e0959-6d7f-401d-b84c-1c098f54bf60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.514246] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5275d83c-b820-621a-9c19-69bf87084e93, 'name': SearchDatastore_Task, 'duration_secs': 0.008187} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.514581] env[61987]: DEBUG oslo_concurrency.lockutils [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.514839] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 89a97f81-bcd9-49e5-bbd3-9140f6638a20/89a97f81-bcd9-49e5-bbd3-9140f6638a20.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 689.515110] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-28c54db1-4c2a-4f3c-b1bf-9f8115983c41 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.521786] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Waiting for the task: (returnval){ [ 689.521786] env[61987]: value = "task-1061282" [ 689.521786] env[61987]: _type = "Task" [ 689.521786] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.529649] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Task: {'id': task-1061282, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.737913] env[61987]: DEBUG oslo_concurrency.lockutils [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Releasing lock "refresh_cache-630e0959-6d7f-401d-b84c-1c098f54bf60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.738295] env[61987]: DEBUG nova.compute.manager [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Instance network_info: |[{"id": "0c62bc7d-9ca1-42c9-b59b-53a18c465220", "address": "fa:16:3e:71:6e:ef", "network": {"id": "c3010fca-d3b0-4624-b84b-b0920f0a0e05", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1462858397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04f95c6e34db49c58e3827d7a72cab54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c62bc7d-9c", "ovs_interfaceid": "0c62bc7d-9ca1-42c9-b59b-53a18c465220", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 689.738603] env[61987]: DEBUG oslo_concurrency.lockutils [req-3e5f6fa6-a6cd-4b9a-8c07-472dfcc32ce0 req-ca57d581-7b9d-40f7-bc75-cf93cad75a00 service nova] Acquired lock "refresh_cache-630e0959-6d7f-401d-b84c-1c098f54bf60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.738782] env[61987]: DEBUG nova.network.neutron [req-3e5f6fa6-a6cd-4b9a-8c07-472dfcc32ce0 req-ca57d581-7b9d-40f7-bc75-cf93cad75a00 service nova] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Refreshing network info cache for port 0c62bc7d-9ca1-42c9-b59b-53a18c465220 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 689.740114] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:6e:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0c62bc7d-9ca1-42c9-b59b-53a18c465220', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 689.747987] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Creating folder: Project (04f95c6e34db49c58e3827d7a72cab54). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 689.748694] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d6d70171-995c-4e4e-b44d-53d1fe9b8341 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.755936] env[61987]: DEBUG nova.scheduler.client.report [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 689.760979] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Created folder: Project (04f95c6e34db49c58e3827d7a72cab54) in parent group-v234219. [ 689.761151] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Creating folder: Instances. Parent ref: group-v234236. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 689.761628] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b38e0a68-9276-4759-abf4-690ccf15731b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.774725] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Created folder: Instances in parent group-v234236. [ 689.774999] env[61987]: DEBUG oslo.service.loopingcall [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 689.775896] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 689.776137] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-979975ba-3230-435a-97ad-aeefd73ce9ae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.797296] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 689.797296] env[61987]: value = "task-1061285" [ 689.797296] env[61987]: _type = "Task" [ 689.797296] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.806260] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061285, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.864879] env[61987]: DEBUG nova.compute.manager [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 689.892935] env[61987]: DEBUG nova.virt.hardware [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 689.893274] env[61987]: DEBUG nova.virt.hardware [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 689.893508] env[61987]: DEBUG nova.virt.hardware [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 689.893709] env[61987]: DEBUG nova.virt.hardware [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 689.893884] env[61987]: DEBUG nova.virt.hardware [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 689.894094] env[61987]: DEBUG nova.virt.hardware [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 689.894375] env[61987]: DEBUG nova.virt.hardware [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 689.894672] env[61987]: DEBUG nova.virt.hardware [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 689.894882] env[61987]: DEBUG nova.virt.hardware [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 689.895052] env[61987]: DEBUG nova.virt.hardware [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 689.895277] env[61987]: DEBUG nova.virt.hardware [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 689.896231] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac8d5548-8c80-42b7-bf78-69513a939cad {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.904908] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99a21557-cbcd-4269-ad9d-3d017e3795fd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.031766] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Task: {'id': task-1061282, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.444288} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.032131] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 89a97f81-bcd9-49e5-bbd3-9140f6638a20/89a97f81-bcd9-49e5-bbd3-9140f6638a20.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 690.032366] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 690.032664] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9e9be5df-6906-4e6a-b97a-ba99671f9c8f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.039278] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Waiting for the task: (returnval){ [ 690.039278] env[61987]: value = "task-1061286" [ 690.039278] env[61987]: _type = "Task" [ 690.039278] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.048184] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Task: {'id': task-1061286, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.262372] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.423s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.262874] env[61987]: DEBUG nova.compute.manager [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 690.265605] env[61987]: DEBUG oslo_concurrency.lockutils [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.442s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.265802] env[61987]: DEBUG oslo_concurrency.lockutils [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.267931] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.419s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.269319] env[61987]: INFO nova.compute.claims [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 690.295997] env[61987]: INFO nova.scheduler.client.report [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Deleted allocations for instance 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59 [ 690.339699] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061285, 'name': CreateVM_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.510722] env[61987]: DEBUG nova.network.neutron [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Successfully updated port: 39a5fbe3-c262-40f4-acdc-48209a7f617a {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 690.550055] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Task: {'id': task-1061286, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.366115} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.550055] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 690.550055] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2334fffc-ef79-488d-895c-24ff429b4797 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.573849] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Reconfiguring VM instance instance-00000020 to attach disk [datastore2] 89a97f81-bcd9-49e5-bbd3-9140f6638a20/89a97f81-bcd9-49e5-bbd3-9140f6638a20.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 690.574145] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4959ae01-241c-421c-a0ce-fa181ab64419 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.589314] env[61987]: DEBUG nova.network.neutron [req-3e5f6fa6-a6cd-4b9a-8c07-472dfcc32ce0 req-ca57d581-7b9d-40f7-bc75-cf93cad75a00 service nova] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Updated VIF entry in instance network info cache for port 0c62bc7d-9ca1-42c9-b59b-53a18c465220. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 690.589622] env[61987]: DEBUG nova.network.neutron [req-3e5f6fa6-a6cd-4b9a-8c07-472dfcc32ce0 req-ca57d581-7b9d-40f7-bc75-cf93cad75a00 service nova] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Updating instance_info_cache with network_info: [{"id": "0c62bc7d-9ca1-42c9-b59b-53a18c465220", "address": "fa:16:3e:71:6e:ef", "network": {"id": "c3010fca-d3b0-4624-b84b-b0920f0a0e05", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1462858397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04f95c6e34db49c58e3827d7a72cab54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c62bc7d-9c", "ovs_interfaceid": "0c62bc7d-9ca1-42c9-b59b-53a18c465220", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.596327] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Waiting for the task: (returnval){ [ 690.596327] env[61987]: value = "task-1061287" [ 690.596327] env[61987]: _type = "Task" [ 690.596327] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.605695] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Task: {'id': task-1061287, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.774168] env[61987]: DEBUG nova.compute.utils [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 690.783332] env[61987]: DEBUG nova.compute.manager [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 690.783332] env[61987]: DEBUG nova.network.neutron [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 690.812756] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061285, 'name': CreateVM_Task, 'duration_secs': 0.826679} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.813447] env[61987]: DEBUG oslo_concurrency.lockutils [None req-be823ad2-bcfb-406b-afcd-f2012f1df25d tempest-TenantUsagesTestJSON-1031022096 tempest-TenantUsagesTestJSON-1031022096-project-member] Lock "6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.496s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.814392] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 690.815144] env[61987]: DEBUG oslo_concurrency.lockutils [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.816658] env[61987]: DEBUG oslo_concurrency.lockutils [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.816658] env[61987]: DEBUG oslo_concurrency.lockutils [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 690.816658] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35689489-df7f-43fa-81bf-b39010002d53 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.822645] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for the task: (returnval){ [ 690.822645] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52773bb9-0144-3c77-b4e3-2fe77bfd5819" [ 690.822645] env[61987]: _type = "Task" [ 690.822645] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.831298] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52773bb9-0144-3c77-b4e3-2fe77bfd5819, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.833489] env[61987]: DEBUG nova.policy [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a3425a6d5e36481bb702ddf789ea9eed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c3d31c138df453486a2b1a6b70a84d4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 691.013529] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Acquiring lock "refresh_cache-466a5f5a-f48a-4921-8bcb-44a1840c8141" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.013700] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Acquired lock "refresh_cache-466a5f5a-f48a-4921-8bcb-44a1840c8141" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.013864] env[61987]: DEBUG nova.network.neutron [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 691.091941] env[61987]: DEBUG oslo_concurrency.lockutils [req-3e5f6fa6-a6cd-4b9a-8c07-472dfcc32ce0 req-ca57d581-7b9d-40f7-bc75-cf93cad75a00 service nova] Releasing lock "refresh_cache-630e0959-6d7f-401d-b84c-1c098f54bf60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.110742] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Task: {'id': task-1061287, 'name': ReconfigVM_Task, 'duration_secs': 0.277596} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.111108] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Reconfigured VM instance instance-00000020 to attach disk [datastore2] 89a97f81-bcd9-49e5-bbd3-9140f6638a20/89a97f81-bcd9-49e5-bbd3-9140f6638a20.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 691.111952] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a191f8b8-343f-4d4c-bce5-76ca4770897d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.118894] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Waiting for the task: (returnval){ [ 691.118894] env[61987]: value = "task-1061288" [ 691.118894] env[61987]: _type = "Task" [ 691.118894] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.129729] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Task: {'id': task-1061288, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.153434] env[61987]: DEBUG nova.network.neutron [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Successfully created port: 574078f6-ac6b-4540-808d-a4b8a1183854 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 691.280923] env[61987]: DEBUG nova.compute.manager [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 691.333896] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52773bb9-0144-3c77-b4e3-2fe77bfd5819, 'name': SearchDatastore_Task, 'duration_secs': 0.010221} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.336458] env[61987]: DEBUG oslo_concurrency.lockutils [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.336697] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 691.336927] env[61987]: DEBUG oslo_concurrency.lockutils [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.337104] env[61987]: DEBUG oslo_concurrency.lockutils [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.337294] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 691.337932] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a2970a3e-3dba-41de-a9ff-8af920c9c69e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.353056] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 691.353056] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 691.353056] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-edb69191-8dad-4170-a69e-f3d61e0a6e53 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.358700] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for the task: (returnval){ [ 691.358700] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52b64656-f1a1-42d1-0806-60245616c74f" [ 691.358700] env[61987]: _type = "Task" [ 691.358700] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.373843] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b64656-f1a1-42d1-0806-60245616c74f, 'name': SearchDatastore_Task, 'duration_secs': 0.008182} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.374660] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98df0235-b903-4c70-a4dc-6662a9176fae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.380029] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for the task: (returnval){ [ 691.380029] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]525859a7-f950-47c2-b214-838f45f4cb08" [ 691.380029] env[61987]: _type = "Task" [ 691.380029] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.394866] env[61987]: DEBUG nova.compute.manager [req-7c191552-e500-4277-918a-be94b7aa3c6d req-b8840d4a-9301-452e-80a8-a8c69048ede6 service nova] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Received event network-vif-plugged-39a5fbe3-c262-40f4-acdc-48209a7f617a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 691.395093] env[61987]: DEBUG oslo_concurrency.lockutils [req-7c191552-e500-4277-918a-be94b7aa3c6d req-b8840d4a-9301-452e-80a8-a8c69048ede6 service nova] Acquiring lock "466a5f5a-f48a-4921-8bcb-44a1840c8141-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.395293] env[61987]: DEBUG oslo_concurrency.lockutils [req-7c191552-e500-4277-918a-be94b7aa3c6d req-b8840d4a-9301-452e-80a8-a8c69048ede6 service nova] Lock "466a5f5a-f48a-4921-8bcb-44a1840c8141-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.395457] env[61987]: DEBUG oslo_concurrency.lockutils [req-7c191552-e500-4277-918a-be94b7aa3c6d req-b8840d4a-9301-452e-80a8-a8c69048ede6 service nova] Lock "466a5f5a-f48a-4921-8bcb-44a1840c8141-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.395618] env[61987]: DEBUG nova.compute.manager [req-7c191552-e500-4277-918a-be94b7aa3c6d req-b8840d4a-9301-452e-80a8-a8c69048ede6 service nova] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] No waiting events found dispatching network-vif-plugged-39a5fbe3-c262-40f4-acdc-48209a7f617a {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 691.395806] env[61987]: WARNING nova.compute.manager [req-7c191552-e500-4277-918a-be94b7aa3c6d req-b8840d4a-9301-452e-80a8-a8c69048ede6 service nova] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Received unexpected event network-vif-plugged-39a5fbe3-c262-40f4-acdc-48209a7f617a for instance with vm_state building and task_state spawning. [ 691.396064] env[61987]: DEBUG nova.compute.manager [req-7c191552-e500-4277-918a-be94b7aa3c6d req-b8840d4a-9301-452e-80a8-a8c69048ede6 service nova] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Received event network-changed-39a5fbe3-c262-40f4-acdc-48209a7f617a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 691.396452] env[61987]: DEBUG nova.compute.manager [req-7c191552-e500-4277-918a-be94b7aa3c6d req-b8840d4a-9301-452e-80a8-a8c69048ede6 service nova] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Refreshing instance network info cache due to event network-changed-39a5fbe3-c262-40f4-acdc-48209a7f617a. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 691.396728] env[61987]: DEBUG oslo_concurrency.lockutils [req-7c191552-e500-4277-918a-be94b7aa3c6d req-b8840d4a-9301-452e-80a8-a8c69048ede6 service nova] Acquiring lock "refresh_cache-466a5f5a-f48a-4921-8bcb-44a1840c8141" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.403785] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]525859a7-f950-47c2-b214-838f45f4cb08, 'name': SearchDatastore_Task, 'duration_secs': 0.009565} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.404017] env[61987]: DEBUG oslo_concurrency.lockutils [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.404269] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 630e0959-6d7f-401d-b84c-1c098f54bf60/630e0959-6d7f-401d-b84c-1c098f54bf60.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 691.404527] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d480b8a7-8ddb-41f1-bb75-2d0b256169ad {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.416649] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for the task: (returnval){ [ 691.416649] env[61987]: value = "task-1061289" [ 691.416649] env[61987]: _type = "Task" [ 691.416649] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.425882] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061289, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.551811] env[61987]: DEBUG nova.network.neutron [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 691.629461] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Task: {'id': task-1061288, 'name': Rename_Task, 'duration_secs': 0.143259} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.630594] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 691.630594] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4c9ca616-90f1-4d74-9180-8b83fb62f687 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.639412] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Waiting for the task: (returnval){ [ 691.639412] env[61987]: value = "task-1061290" [ 691.639412] env[61987]: _type = "Task" [ 691.639412] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.655029] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Task: {'id': task-1061290, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.678772] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-123014a6-e35c-4180-91c4-adc6a0176ead {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.689145] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b29a59c4-8f63-49b4-904d-b041a38c9741 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.724271] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2180d4e0-b356-4ea8-9fe2-0974d422bb04 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.732270] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9544ec73-da43-4f33-9557-aba1bd649f8f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.746129] env[61987]: DEBUG nova.compute.provider_tree [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.748210] env[61987]: DEBUG nova.network.neutron [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Updating instance_info_cache with network_info: [{"id": "39a5fbe3-c262-40f4-acdc-48209a7f617a", "address": "fa:16:3e:f9:43:9a", "network": {"id": "020556c5-6e13-45fa-a715-bf097bdf0f73", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-367382123-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3adce5cde814bf0bb57b8e941dba79f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39a5fbe3-c2", "ovs_interfaceid": "39a5fbe3-c262-40f4-acdc-48209a7f617a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.928746] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061289, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502412} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.929069] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 630e0959-6d7f-401d-b84c-1c098f54bf60/630e0959-6d7f-401d-b84c-1c098f54bf60.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 691.929320] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 691.929708] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-65a531cb-9e22-426b-b29d-ba3777ebf0b7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.935729] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for the task: (returnval){ [ 691.935729] env[61987]: value = "task-1061291" [ 691.935729] env[61987]: _type = "Task" [ 691.935729] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.944988] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061291, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.155669] env[61987]: DEBUG oslo_vmware.api [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Task: {'id': task-1061290, 'name': PowerOnVM_Task, 'duration_secs': 0.512825} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.156867] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 692.157121] env[61987]: INFO nova.compute.manager [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Took 7.25 seconds to spawn the instance on the hypervisor. [ 692.157305] env[61987]: DEBUG nova.compute.manager [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 692.158335] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c8a5570-fd45-4b75-beb0-737c1a5b0a32 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.250764] env[61987]: DEBUG nova.scheduler.client.report [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 692.254043] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Releasing lock "refresh_cache-466a5f5a-f48a-4921-8bcb-44a1840c8141" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.254908] env[61987]: DEBUG nova.compute.manager [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Instance network_info: |[{"id": "39a5fbe3-c262-40f4-acdc-48209a7f617a", "address": "fa:16:3e:f9:43:9a", "network": {"id": "020556c5-6e13-45fa-a715-bf097bdf0f73", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-367382123-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3adce5cde814bf0bb57b8e941dba79f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39a5fbe3-c2", "ovs_interfaceid": "39a5fbe3-c262-40f4-acdc-48209a7f617a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 692.255969] env[61987]: DEBUG oslo_concurrency.lockutils [req-7c191552-e500-4277-918a-be94b7aa3c6d req-b8840d4a-9301-452e-80a8-a8c69048ede6 service nova] Acquired lock "refresh_cache-466a5f5a-f48a-4921-8bcb-44a1840c8141" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.255969] env[61987]: DEBUG nova.network.neutron [req-7c191552-e500-4277-918a-be94b7aa3c6d req-b8840d4a-9301-452e-80a8-a8c69048ede6 service nova] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Refreshing network info cache for port 39a5fbe3-c262-40f4-acdc-48209a7f617a {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 692.256382] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f9:43:9a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56834f67-27a8-43dc-bbc6-a74aaa08959b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '39a5fbe3-c262-40f4-acdc-48209a7f617a', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 692.264807] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Creating folder: Project (d3adce5cde814bf0bb57b8e941dba79f). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 692.265673] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6628c35e-7ed6-4308-8dca-001e88da61d8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.276857] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Created folder: Project (d3adce5cde814bf0bb57b8e941dba79f) in parent group-v234219. [ 692.277057] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Creating folder: Instances. Parent ref: group-v234239. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 692.277289] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-08ebae4e-00ee-4bcb-8f04-82361633c6a5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.286581] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Created folder: Instances in parent group-v234239. [ 692.286818] env[61987]: DEBUG oslo.service.loopingcall [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 692.287013] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 692.287218] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-623ceea8-2534-4b45-9a14-703c94b6a682 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.303780] env[61987]: DEBUG nova.compute.manager [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 692.310499] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 692.310499] env[61987]: value = "task-1061294" [ 692.310499] env[61987]: _type = "Task" [ 692.310499] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.318074] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061294, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.332999] env[61987]: DEBUG nova.virt.hardware [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 692.332999] env[61987]: DEBUG nova.virt.hardware [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 692.332999] env[61987]: DEBUG nova.virt.hardware [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 692.333222] env[61987]: DEBUG nova.virt.hardware [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 692.333222] env[61987]: DEBUG nova.virt.hardware [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 692.333222] env[61987]: DEBUG nova.virt.hardware [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 692.333478] env[61987]: DEBUG nova.virt.hardware [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 692.334767] env[61987]: DEBUG nova.virt.hardware [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 692.334767] env[61987]: DEBUG nova.virt.hardware [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 692.334767] env[61987]: DEBUG nova.virt.hardware [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 692.334767] env[61987]: DEBUG nova.virt.hardware [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 692.335078] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c49b2907-9ba6-4e1b-956f-a87f5e66f398 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.343181] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e2c9dcd-b0b2-45e9-9565-898899942ce0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.445191] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061291, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076188} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.445365] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 692.446127] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f58002f-211f-434e-9d82-d9702ffefe12 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.476812] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Reconfiguring VM instance instance-00000021 to attach disk [datastore2] 630e0959-6d7f-401d-b84c-1c098f54bf60/630e0959-6d7f-401d-b84c-1c098f54bf60.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 692.476812] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71072b3f-ace1-44bc-a274-61b1047ee4c4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.496140] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for the task: (returnval){ [ 692.496140] env[61987]: value = "task-1061295" [ 692.496140] env[61987]: _type = "Task" [ 692.496140] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.508027] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061295, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.678405] env[61987]: INFO nova.compute.manager [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Took 37.65 seconds to build instance. [ 692.756212] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.488s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.756836] env[61987]: DEBUG nova.compute.manager [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 692.759608] env[61987]: DEBUG oslo_concurrency.lockutils [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.900s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.761161] env[61987]: INFO nova.compute.claims [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 692.779377] env[61987]: DEBUG nova.network.neutron [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Successfully updated port: 574078f6-ac6b-4540-808d-a4b8a1183854 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 692.827532] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061294, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.006197] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061295, 'name': ReconfigVM_Task, 'duration_secs': 0.495139} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.007171] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Reconfigured VM instance instance-00000021 to attach disk [datastore2] 630e0959-6d7f-401d-b84c-1c098f54bf60/630e0959-6d7f-401d-b84c-1c098f54bf60.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 693.007746] env[61987]: DEBUG nova.network.neutron [req-7c191552-e500-4277-918a-be94b7aa3c6d req-b8840d4a-9301-452e-80a8-a8c69048ede6 service nova] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Updated VIF entry in instance network info cache for port 39a5fbe3-c262-40f4-acdc-48209a7f617a. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 693.008578] env[61987]: DEBUG nova.network.neutron [req-7c191552-e500-4277-918a-be94b7aa3c6d req-b8840d4a-9301-452e-80a8-a8c69048ede6 service nova] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Updating instance_info_cache with network_info: [{"id": "39a5fbe3-c262-40f4-acdc-48209a7f617a", "address": "fa:16:3e:f9:43:9a", "network": {"id": "020556c5-6e13-45fa-a715-bf097bdf0f73", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-367382123-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3adce5cde814bf0bb57b8e941dba79f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39a5fbe3-c2", "ovs_interfaceid": "39a5fbe3-c262-40f4-acdc-48209a7f617a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.009237] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3c036462-6468-44a1-a638-22bf834f2871 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.015083] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for the task: (returnval){ [ 693.015083] env[61987]: value = "task-1061296" [ 693.015083] env[61987]: _type = "Task" [ 693.015083] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.025189] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061296, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.179629] env[61987]: DEBUG oslo_concurrency.lockutils [None req-47a46587-c34c-46e5-9cfd-a3a6b3fa18ff tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Lock "89a97f81-bcd9-49e5-bbd3-9140f6638a20" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.678s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.267399] env[61987]: DEBUG nova.compute.utils [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 693.268718] env[61987]: DEBUG nova.compute.manager [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 693.268892] env[61987]: DEBUG nova.network.neutron [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 693.281722] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "refresh_cache-ea6c2120-a69a-42c7-bc92-3e74539c008d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.281863] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquired lock "refresh_cache-ea6c2120-a69a-42c7-bc92-3e74539c008d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.282012] env[61987]: DEBUG nova.network.neutron [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 693.308040] env[61987]: DEBUG nova.policy [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3011f049be54b2cb8922d646b94a310', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd828262de7b14d2ab38193d7d34e8f7a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 693.320608] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061294, 'name': CreateVM_Task, 'duration_secs': 0.550093} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.320773] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 693.321612] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.321699] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.321934] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 693.322197] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36fecddd-039d-493a-9597-6bd9fe348ff5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.327091] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Waiting for the task: (returnval){ [ 693.327091] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]521121fb-7403-95f5-a1bf-f92f52980447" [ 693.327091] env[61987]: _type = "Task" [ 693.327091] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.335264] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]521121fb-7403-95f5-a1bf-f92f52980447, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.509828] env[61987]: DEBUG nova.compute.manager [req-1e7f27cb-7ed3-4625-b0ad-bfce03ceb400 req-52de8852-79e0-4239-bf88-e0acfb6f4d99 service nova] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Received event network-vif-plugged-574078f6-ac6b-4540-808d-a4b8a1183854 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 693.509828] env[61987]: DEBUG oslo_concurrency.lockutils [req-1e7f27cb-7ed3-4625-b0ad-bfce03ceb400 req-52de8852-79e0-4239-bf88-e0acfb6f4d99 service nova] Acquiring lock "ea6c2120-a69a-42c7-bc92-3e74539c008d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.509828] env[61987]: DEBUG oslo_concurrency.lockutils [req-1e7f27cb-7ed3-4625-b0ad-bfce03ceb400 req-52de8852-79e0-4239-bf88-e0acfb6f4d99 service nova] Lock "ea6c2120-a69a-42c7-bc92-3e74539c008d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.509828] env[61987]: DEBUG oslo_concurrency.lockutils [req-1e7f27cb-7ed3-4625-b0ad-bfce03ceb400 req-52de8852-79e0-4239-bf88-e0acfb6f4d99 service nova] Lock "ea6c2120-a69a-42c7-bc92-3e74539c008d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.509828] env[61987]: DEBUG nova.compute.manager [req-1e7f27cb-7ed3-4625-b0ad-bfce03ceb400 req-52de8852-79e0-4239-bf88-e0acfb6f4d99 service nova] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] No waiting events found dispatching network-vif-plugged-574078f6-ac6b-4540-808d-a4b8a1183854 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 693.510373] env[61987]: WARNING nova.compute.manager [req-1e7f27cb-7ed3-4625-b0ad-bfce03ceb400 req-52de8852-79e0-4239-bf88-e0acfb6f4d99 service nova] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Received unexpected event network-vif-plugged-574078f6-ac6b-4540-808d-a4b8a1183854 for instance with vm_state building and task_state spawning. [ 693.510373] env[61987]: DEBUG nova.compute.manager [req-1e7f27cb-7ed3-4625-b0ad-bfce03ceb400 req-52de8852-79e0-4239-bf88-e0acfb6f4d99 service nova] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Received event network-changed-574078f6-ac6b-4540-808d-a4b8a1183854 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 693.510373] env[61987]: DEBUG nova.compute.manager [req-1e7f27cb-7ed3-4625-b0ad-bfce03ceb400 req-52de8852-79e0-4239-bf88-e0acfb6f4d99 service nova] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Refreshing instance network info cache due to event network-changed-574078f6-ac6b-4540-808d-a4b8a1183854. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 693.510373] env[61987]: DEBUG oslo_concurrency.lockutils [req-1e7f27cb-7ed3-4625-b0ad-bfce03ceb400 req-52de8852-79e0-4239-bf88-e0acfb6f4d99 service nova] Acquiring lock "refresh_cache-ea6c2120-a69a-42c7-bc92-3e74539c008d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.512365] env[61987]: DEBUG oslo_concurrency.lockutils [req-7c191552-e500-4277-918a-be94b7aa3c6d req-b8840d4a-9301-452e-80a8-a8c69048ede6 service nova] Releasing lock "refresh_cache-466a5f5a-f48a-4921-8bcb-44a1840c8141" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.525733] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061296, 'name': Rename_Task, 'duration_secs': 0.126187} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.525994] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 693.526271] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-621470fc-05f3-4419-80e6-6f356e46bc66 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.533442] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for the task: (returnval){ [ 693.533442] env[61987]: value = "task-1061297" [ 693.533442] env[61987]: _type = "Task" [ 693.533442] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.546041] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061297, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.656853] env[61987]: DEBUG nova.network.neutron [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Successfully created port: a55b478b-de12-47cb-80e8-07a467463422 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 693.682247] env[61987]: DEBUG nova.compute.manager [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 693.772359] env[61987]: DEBUG nova.compute.manager [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 693.828454] env[61987]: DEBUG nova.network.neutron [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 693.841600] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]521121fb-7403-95f5-a1bf-f92f52980447, 'name': SearchDatastore_Task, 'duration_secs': 0.009681} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.841889] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.842133] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 693.842369] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.842526] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.842781] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 693.843047] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fa9ee583-f6be-4406-85da-da79b66a681f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.855008] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 693.855194] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 693.855921] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd762667-0f94-40c6-a531-0972f2fe2010 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.867447] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Waiting for the task: (returnval){ [ 693.867447] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52323f93-fabd-2b6d-0f69-0f75daddacc1" [ 693.867447] env[61987]: _type = "Task" [ 693.867447] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.875855] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52323f93-fabd-2b6d-0f69-0f75daddacc1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.046692] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061297, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.166517] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f133d0cb-289a-4262-bae5-36ee4e4aceb9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.173788] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1409aca4-ce5f-4e42-a9ab-2ec4d6f44b64 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.211164] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-646a9a69-9f02-4ef3-8bfd-483d7e48d111 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.223326] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9137eb6c-188f-447e-b1bb-71310f00782f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.235611] env[61987]: DEBUG nova.compute.provider_tree [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 694.238079] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.273791] env[61987]: DEBUG nova.network.neutron [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Updating instance_info_cache with network_info: [{"id": "574078f6-ac6b-4540-808d-a4b8a1183854", "address": "fa:16:3e:5e:7b:09", "network": {"id": "8dfc473a-6d57-4169-9338-49a9565bb93d", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-422412073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c3d31c138df453486a2b1a6b70a84d4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap574078f6-ac", "ovs_interfaceid": "574078f6-ac6b-4540-808d-a4b8a1183854", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.377089] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52323f93-fabd-2b6d-0f69-0f75daddacc1, 'name': SearchDatastore_Task, 'duration_secs': 0.008263} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.377869] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3406fb8c-fe01-44de-9b3a-f3bfb092f365 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.382986] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Waiting for the task: (returnval){ [ 694.382986] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]524ba220-5647-dd85-87ae-7557af3d91df" [ 694.382986] env[61987]: _type = "Task" [ 694.382986] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.391217] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524ba220-5647-dd85-87ae-7557af3d91df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.546679] env[61987]: DEBUG oslo_vmware.api [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061297, 'name': PowerOnVM_Task, 'duration_secs': 0.532136} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.546679] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 694.546679] env[61987]: INFO nova.compute.manager [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Took 7.18 seconds to spawn the instance on the hypervisor. [ 694.546904] env[61987]: DEBUG nova.compute.manager [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 694.547636] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63fa5435-8c93-4777-8dbc-9ea25c5e7454 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.739699] env[61987]: DEBUG nova.scheduler.client.report [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 694.776048] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Releasing lock "refresh_cache-ea6c2120-a69a-42c7-bc92-3e74539c008d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.776339] env[61987]: DEBUG nova.compute.manager [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Instance network_info: |[{"id": "574078f6-ac6b-4540-808d-a4b8a1183854", "address": "fa:16:3e:5e:7b:09", "network": {"id": "8dfc473a-6d57-4169-9338-49a9565bb93d", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-422412073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c3d31c138df453486a2b1a6b70a84d4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap574078f6-ac", "ovs_interfaceid": "574078f6-ac6b-4540-808d-a4b8a1183854", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 694.776681] env[61987]: DEBUG oslo_concurrency.lockutils [req-1e7f27cb-7ed3-4625-b0ad-bfce03ceb400 req-52de8852-79e0-4239-bf88-e0acfb6f4d99 service nova] Acquired lock "refresh_cache-ea6c2120-a69a-42c7-bc92-3e74539c008d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.777188] env[61987]: DEBUG nova.network.neutron [req-1e7f27cb-7ed3-4625-b0ad-bfce03ceb400 req-52de8852-79e0-4239-bf88-e0acfb6f4d99 service nova] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Refreshing network info cache for port 574078f6-ac6b-4540-808d-a4b8a1183854 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 694.778495] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:7b:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae4e3171-21cd-4094-b6cf-81bf366c75bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '574078f6-ac6b-4540-808d-a4b8a1183854', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 694.786704] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Creating folder: Project (0c3d31c138df453486a2b1a6b70a84d4). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 694.787452] env[61987]: DEBUG nova.compute.manager [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 694.789825] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6d66465e-40df-4b39-b022-668c4c3009da {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.803097] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Created folder: Project (0c3d31c138df453486a2b1a6b70a84d4) in parent group-v234219. [ 694.803237] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Creating folder: Instances. Parent ref: group-v234242. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 694.803578] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d1da0530-6984-4b0c-81bf-51e6b2239119 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.812594] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Created folder: Instances in parent group-v234242. [ 694.813188] env[61987]: DEBUG oslo.service.loopingcall [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 694.813391] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 694.814024] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7767d00d-6b07-4ecd-b6ad-f09d08c31ab3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.830913] env[61987]: DEBUG nova.virt.hardware [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 694.831181] env[61987]: DEBUG nova.virt.hardware [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 694.831361] env[61987]: DEBUG nova.virt.hardware [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 694.831577] env[61987]: DEBUG nova.virt.hardware [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 694.831730] env[61987]: DEBUG nova.virt.hardware [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 694.831880] env[61987]: DEBUG nova.virt.hardware [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 694.832096] env[61987]: DEBUG nova.virt.hardware [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 694.832260] env[61987]: DEBUG nova.virt.hardware [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 694.832445] env[61987]: DEBUG nova.virt.hardware [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 694.832612] env[61987]: DEBUG nova.virt.hardware [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 694.832781] env[61987]: DEBUG nova.virt.hardware [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 694.833528] env[61987]: DEBUG oslo_concurrency.lockutils [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Acquiring lock "89a97f81-bcd9-49e5-bbd3-9140f6638a20" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.833751] env[61987]: DEBUG oslo_concurrency.lockutils [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Lock "89a97f81-bcd9-49e5-bbd3-9140f6638a20" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.833945] env[61987]: DEBUG oslo_concurrency.lockutils [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Acquiring lock "89a97f81-bcd9-49e5-bbd3-9140f6638a20-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.834164] env[61987]: DEBUG oslo_concurrency.lockutils [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Lock "89a97f81-bcd9-49e5-bbd3-9140f6638a20-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.834339] env[61987]: DEBUG oslo_concurrency.lockutils [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Lock "89a97f81-bcd9-49e5-bbd3-9140f6638a20-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.836486] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09987749-4ad7-4f60-acd6-2559feb462a9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.839588] env[61987]: INFO nova.compute.manager [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Terminating instance [ 694.848994] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a07560f1-927f-4d48-a088-2e0c9e5f09ec {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.851658] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 694.851658] env[61987]: value = "task-1061300" [ 694.851658] env[61987]: _type = "Task" [ 694.851658] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.869476] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061300, 'name': CreateVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.892729] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524ba220-5647-dd85-87ae-7557af3d91df, 'name': SearchDatastore_Task, 'duration_secs': 0.008533} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.892966] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.893302] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 466a5f5a-f48a-4921-8bcb-44a1840c8141/466a5f5a-f48a-4921-8bcb-44a1840c8141.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 694.893576] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-140e4578-c656-4bfe-a641-6bab5d352475 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.903176] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Waiting for the task: (returnval){ [ 694.903176] env[61987]: value = "task-1061301" [ 694.903176] env[61987]: _type = "Task" [ 694.903176] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.912288] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': task-1061301, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.065194] env[61987]: INFO nova.compute.manager [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Took 38.10 seconds to build instance. [ 695.246454] env[61987]: DEBUG oslo_concurrency.lockutils [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.486s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.246791] env[61987]: DEBUG nova.compute.manager [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 695.250044] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.273s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.251758] env[61987]: INFO nova.compute.claims [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 695.344020] env[61987]: DEBUG nova.compute.manager [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 695.344020] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 695.344989] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba464f9d-4f44-444e-ad99-71c26ef5e57d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.359933] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 695.361178] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d8ca824d-c5b0-4ae7-b09a-5d81e5eafb14 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.368920] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061300, 'name': CreateVM_Task, 'duration_secs': 0.351184} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.370324] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 695.370976] env[61987]: DEBUG oslo_vmware.api [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Waiting for the task: (returnval){ [ 695.370976] env[61987]: value = "task-1061302" [ 695.370976] env[61987]: _type = "Task" [ 695.370976] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.372787] env[61987]: DEBUG oslo_vmware.service [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a5c397-b60a-411d-b41e-e4f1eef1644e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.384165] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.384600] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.385420] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 695.385862] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97ceb63a-76ac-47ae-8dde-becb60069f88 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.392014] env[61987]: DEBUG oslo_vmware.api [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Task: {'id': task-1061302, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.396240] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 695.396240] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52e7726a-ef0d-4748-26b7-1eba8582c6b3" [ 695.396240] env[61987]: _type = "Task" [ 695.396240] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.406537] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52e7726a-ef0d-4748-26b7-1eba8582c6b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.414801] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': task-1061301, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.553564] env[61987]: DEBUG nova.compute.manager [req-f9a84d4b-690e-4f41-b2b6-8ebbb13f4023 req-8aa786ca-28a8-4c67-858b-43797d2821f2 service nova] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Received event network-vif-plugged-a55b478b-de12-47cb-80e8-07a467463422 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 695.553855] env[61987]: DEBUG oslo_concurrency.lockutils [req-f9a84d4b-690e-4f41-b2b6-8ebbb13f4023 req-8aa786ca-28a8-4c67-858b-43797d2821f2 service nova] Acquiring lock "02fe490b-75c8-406b-98e0-aa200f249673-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.554085] env[61987]: DEBUG oslo_concurrency.lockutils [req-f9a84d4b-690e-4f41-b2b6-8ebbb13f4023 req-8aa786ca-28a8-4c67-858b-43797d2821f2 service nova] Lock "02fe490b-75c8-406b-98e0-aa200f249673-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.554266] env[61987]: DEBUG oslo_concurrency.lockutils [req-f9a84d4b-690e-4f41-b2b6-8ebbb13f4023 req-8aa786ca-28a8-4c67-858b-43797d2821f2 service nova] Lock "02fe490b-75c8-406b-98e0-aa200f249673-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.554406] env[61987]: DEBUG nova.compute.manager [req-f9a84d4b-690e-4f41-b2b6-8ebbb13f4023 req-8aa786ca-28a8-4c67-858b-43797d2821f2 service nova] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] No waiting events found dispatching network-vif-plugged-a55b478b-de12-47cb-80e8-07a467463422 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 695.554593] env[61987]: WARNING nova.compute.manager [req-f9a84d4b-690e-4f41-b2b6-8ebbb13f4023 req-8aa786ca-28a8-4c67-858b-43797d2821f2 service nova] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Received unexpected event network-vif-plugged-a55b478b-de12-47cb-80e8-07a467463422 for instance with vm_state building and task_state spawning. [ 695.567681] env[61987]: DEBUG oslo_concurrency.lockutils [None req-007453be-9a2b-481f-b950-3bb4aac64cff tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Lock "630e0959-6d7f-401d-b84c-1c098f54bf60" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 123.955s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.670863] env[61987]: DEBUG nova.network.neutron [req-1e7f27cb-7ed3-4625-b0ad-bfce03ceb400 req-52de8852-79e0-4239-bf88-e0acfb6f4d99 service nova] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Updated VIF entry in instance network info cache for port 574078f6-ac6b-4540-808d-a4b8a1183854. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 695.671234] env[61987]: DEBUG nova.network.neutron [req-1e7f27cb-7ed3-4625-b0ad-bfce03ceb400 req-52de8852-79e0-4239-bf88-e0acfb6f4d99 service nova] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Updating instance_info_cache with network_info: [{"id": "574078f6-ac6b-4540-808d-a4b8a1183854", "address": "fa:16:3e:5e:7b:09", "network": {"id": "8dfc473a-6d57-4169-9338-49a9565bb93d", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-422412073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c3d31c138df453486a2b1a6b70a84d4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap574078f6-ac", "ovs_interfaceid": "574078f6-ac6b-4540-808d-a4b8a1183854", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.706131] env[61987]: DEBUG nova.network.neutron [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Successfully updated port: a55b478b-de12-47cb-80e8-07a467463422 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 695.756309] env[61987]: DEBUG nova.compute.utils [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 695.759895] env[61987]: DEBUG nova.compute.manager [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 695.760099] env[61987]: DEBUG nova.network.neutron [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 695.798089] env[61987]: DEBUG nova.policy [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f8261038f3a446e688eac1804d4975fe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c062bbfea30a42bbb2dbec5a84358d95', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 695.885052] env[61987]: DEBUG oslo_vmware.api [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Task: {'id': task-1061302, 'name': PowerOffVM_Task, 'duration_secs': 0.2518} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.885300] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 695.885472] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 695.885722] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b1934115-685a-41b6-adc4-6279ff3c62c9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.906344] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.906715] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 695.906828] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.906977] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.907170] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 695.907488] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-44ea848c-1245-483d-907c-881c7340b3ca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.920858] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 695.921148] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 695.922386] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6172e3ba-2b74-422a-ae52-0f31acdfd187 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.930403] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': task-1061301, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.532926} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.931049] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 466a5f5a-f48a-4921-8bcb-44a1840c8141/466a5f5a-f48a-4921-8bcb-44a1840c8141.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 695.931793] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 695.931793] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-415c9d57-6295-424a-88d2-bae803133060 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.935820] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ef8c717-0b5c-4cc2-8cec-8cff4724c319 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.940282] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Waiting for the task: (returnval){ [ 695.940282] env[61987]: value = "task-1061304" [ 695.940282] env[61987]: _type = "Task" [ 695.940282] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.941696] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 695.941696] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5271b6fe-ad7f-ad7a-ac5c-76358f258921" [ 695.941696] env[61987]: _type = "Task" [ 695.941696] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.952496] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': task-1061304, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.955650] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5271b6fe-ad7f-ad7a-ac5c-76358f258921, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.072852] env[61987]: DEBUG nova.compute.manager [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 696.115385] env[61987]: DEBUG nova.network.neutron [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Successfully created port: 2b8e0044-50eb-40c8-8a96-9c828626d2f7 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 696.136246] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 696.136246] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 696.136379] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Deleting the datastore file [datastore2] 89a97f81-bcd9-49e5-bbd3-9140f6638a20 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 696.136604] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0a1123e7-e239-4fe3-a4c7-406b6da9ef92 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.145027] env[61987]: DEBUG oslo_vmware.api [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Waiting for the task: (returnval){ [ 696.145027] env[61987]: value = "task-1061305" [ 696.145027] env[61987]: _type = "Task" [ 696.145027] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.151019] env[61987]: DEBUG oslo_vmware.api [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Task: {'id': task-1061305, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.174012] env[61987]: DEBUG oslo_concurrency.lockutils [req-1e7f27cb-7ed3-4625-b0ad-bfce03ceb400 req-52de8852-79e0-4239-bf88-e0acfb6f4d99 service nova] Releasing lock "refresh_cache-ea6c2120-a69a-42c7-bc92-3e74539c008d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.209045] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "refresh_cache-02fe490b-75c8-406b-98e0-aa200f249673" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.209175] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquired lock "refresh_cache-02fe490b-75c8-406b-98e0-aa200f249673" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.209335] env[61987]: DEBUG nova.network.neutron [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 696.261818] env[61987]: DEBUG nova.compute.manager [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 696.455146] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': task-1061304, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.092956} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.463022] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 696.463022] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-249c6711-63b0-47b5-acd9-7a017251ade0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.465752] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Preparing fetch location {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 696.467118] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Creating directory with path [datastore1] vmware_temp/e3eaa15d-e0db-416b-af95-91e76c2b660e/7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 696.468720] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f5fbe4b3-46cb-43c3-9c61-a7b5cfaeef9f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.493313] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Reconfiguring VM instance instance-00000022 to attach disk [datastore2] 466a5f5a-f48a-4921-8bcb-44a1840c8141/466a5f5a-f48a-4921-8bcb-44a1840c8141.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 696.497345] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e848350e-c3f2-40de-86e5-b41572f9cece {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.512129] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Created directory with path [datastore1] vmware_temp/e3eaa15d-e0db-416b-af95-91e76c2b660e/7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 696.512364] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Fetch image to [datastore1] vmware_temp/e3eaa15d-e0db-416b-af95-91e76c2b660e/7bf8969f-d179-42af-9e15-527d5449f929/tmp-sparse.vmdk {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 696.512539] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Downloading image file data 7bf8969f-d179-42af-9e15-527d5449f929 to [datastore1] vmware_temp/e3eaa15d-e0db-416b-af95-91e76c2b660e/7bf8969f-d179-42af-9e15-527d5449f929/tmp-sparse.vmdk on the data store datastore1 {{(pid=61987) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 696.513674] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc66b382-abe9-4657-8472-8619f2ae55b3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.517437] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Waiting for the task: (returnval){ [ 696.517437] env[61987]: value = "task-1061306" [ 696.517437] env[61987]: _type = "Task" [ 696.517437] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.526580] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': task-1061306, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.529188] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75fcd3c2-396b-42e5-935c-3090dbbeecaa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.542891] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b649c5a9-5532-4871-9cff-749a5edad4e8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.588289] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96b80b10-d49b-4ad8-87ee-ec629fee4ba4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.595811] env[61987]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-8762734e-6ae1-478a-8b51-eedd844fe2cf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.605520] env[61987]: DEBUG oslo_concurrency.lockutils [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.618147] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Downloading image file data 7bf8969f-d179-42af-9e15-527d5449f929 to the data store datastore1 {{(pid=61987) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 696.655563] env[61987]: DEBUG oslo_vmware.api [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Task: {'id': task-1061305, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.462296} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.658138] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 696.658340] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 696.658518] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 696.658693] env[61987]: INFO nova.compute.manager [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Took 1.31 seconds to destroy the instance on the hypervisor. [ 696.658936] env[61987]: DEBUG oslo.service.loopingcall [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 696.659319] env[61987]: DEBUG nova.compute.manager [-] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 696.659422] env[61987]: DEBUG nova.network.neutron [-] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 696.679043] env[61987]: DEBUG oslo_vmware.rw_handles [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/e3eaa15d-e0db-416b-af95-91e76c2b660e/7bf8969f-d179-42af-9e15-527d5449f929/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61987) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 696.777502] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-766d0607-c42a-40a4-b7cd-ec4c9bc6973e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.792020] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52681ee5-c48e-47ab-afac-b93386cc69aa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.829860] env[61987]: DEBUG nova.network.neutron [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.832385] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ee87d7-59c6-4f7a-8993-743964c24379 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.845475] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc7121e5-701d-4c74-a1e2-34b3bac3a3a5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.860253] env[61987]: DEBUG nova.compute.provider_tree [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 697.033643] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': task-1061306, 'name': ReconfigVM_Task, 'duration_secs': 0.502467} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.033643] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Reconfigured VM instance instance-00000022 to attach disk [datastore2] 466a5f5a-f48a-4921-8bcb-44a1840c8141/466a5f5a-f48a-4921-8bcb-44a1840c8141.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 697.034260] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d45ee302-d309-4533-b8a2-6918c29d5433 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.042679] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Waiting for the task: (returnval){ [ 697.042679] env[61987]: value = "task-1061307" [ 697.042679] env[61987]: _type = "Task" [ 697.042679] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.056024] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': task-1061307, 'name': Rename_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.085120] env[61987]: DEBUG nova.compute.manager [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 697.086100] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32371156-f076-4ac2-9ea1-1eab28954356 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.092151] env[61987]: DEBUG nova.network.neutron [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Updating instance_info_cache with network_info: [{"id": "a55b478b-de12-47cb-80e8-07a467463422", "address": "fa:16:3e:b5:73:24", "network": {"id": "e2d0ce90-ff14-47de-bc4e-a2564d1268a7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-222691221-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d828262de7b14d2ab38193d7d34e8f7a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa55b478b-de", "ovs_interfaceid": "a55b478b-de12-47cb-80e8-07a467463422", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.276755] env[61987]: DEBUG nova.compute.manager [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 697.301059] env[61987]: DEBUG nova.virt.hardware [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 697.301402] env[61987]: DEBUG nova.virt.hardware [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 697.301609] env[61987]: DEBUG nova.virt.hardware [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 697.301815] env[61987]: DEBUG nova.virt.hardware [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 697.301988] env[61987]: DEBUG nova.virt.hardware [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 697.302162] env[61987]: DEBUG nova.virt.hardware [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 697.302391] env[61987]: DEBUG nova.virt.hardware [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 697.302563] env[61987]: DEBUG nova.virt.hardware [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 697.302732] env[61987]: DEBUG nova.virt.hardware [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 697.302927] env[61987]: DEBUG nova.virt.hardware [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 697.303147] env[61987]: DEBUG nova.virt.hardware [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 697.304410] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-475d14c8-081d-4365-afed-c54a1d193b0c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.312216] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67642e6f-d205-44a6-92d7-8f15d03d5f5e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.363985] env[61987]: DEBUG nova.scheduler.client.report [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 697.441349] env[61987]: DEBUG oslo_vmware.rw_handles [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Completed reading data from the image iterator. {{(pid=61987) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 697.441823] env[61987]: DEBUG oslo_vmware.rw_handles [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/e3eaa15d-e0db-416b-af95-91e76c2b660e/7bf8969f-d179-42af-9e15-527d5449f929/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61987) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 697.502839] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Downloaded image file data 7bf8969f-d179-42af-9e15-527d5449f929 to vmware_temp/e3eaa15d-e0db-416b-af95-91e76c2b660e/7bf8969f-d179-42af-9e15-527d5449f929/tmp-sparse.vmdk on the data store datastore1 {{(pid=61987) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 697.505167] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Caching image {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 697.505424] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Copying Virtual Disk [datastore1] vmware_temp/e3eaa15d-e0db-416b-af95-91e76c2b660e/7bf8969f-d179-42af-9e15-527d5449f929/tmp-sparse.vmdk to [datastore1] vmware_temp/e3eaa15d-e0db-416b-af95-91e76c2b660e/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 697.505708] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5ec72651-5ff1-4240-912d-409663d5322d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.514233] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 697.514233] env[61987]: value = "task-1061308" [ 697.514233] env[61987]: _type = "Task" [ 697.514233] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.523686] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061308, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.551219] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': task-1061307, 'name': Rename_Task, 'duration_secs': 0.276222} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.551522] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 697.551774] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ae6649a5-7289-4db2-94db-aa22d9666ff8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.558635] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Waiting for the task: (returnval){ [ 697.558635] env[61987]: value = "task-1061309" [ 697.558635] env[61987]: _type = "Task" [ 697.558635] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.566987] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': task-1061309, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.594714] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Releasing lock "refresh_cache-02fe490b-75c8-406b-98e0-aa200f249673" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.595064] env[61987]: DEBUG nova.compute.manager [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Instance network_info: |[{"id": "a55b478b-de12-47cb-80e8-07a467463422", "address": "fa:16:3e:b5:73:24", "network": {"id": "e2d0ce90-ff14-47de-bc4e-a2564d1268a7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-222691221-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d828262de7b14d2ab38193d7d34e8f7a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa55b478b-de", "ovs_interfaceid": "a55b478b-de12-47cb-80e8-07a467463422", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 697.595527] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b5:73:24', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f52a458-d157-48a3-b4e2-b8cc0779afe2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a55b478b-de12-47cb-80e8-07a467463422', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 697.604103] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Creating folder: Project (d828262de7b14d2ab38193d7d34e8f7a). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 697.604360] env[61987]: INFO nova.compute.manager [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] instance snapshotting [ 697.605790] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-23b69ab0-2f8c-409f-933e-63f978a5d7f3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.608362] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-541f6262-7c0b-48d5-a4df-e98115967935 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.631999] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b4d85c5-e088-4928-b722-ec18c20fac1a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.634935] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Created folder: Project (d828262de7b14d2ab38193d7d34e8f7a) in parent group-v234219. [ 697.635147] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Creating folder: Instances. Parent ref: group-v234245. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 697.635394] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7bc6a8b1-c01c-47c7-9b53-a16a8ee21dea {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.645093] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Created folder: Instances in parent group-v234245. [ 697.645418] env[61987]: DEBUG oslo.service.loopingcall [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 697.645621] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 697.645825] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-be4b108b-a50e-4588-af9e-ea39cd11c4e1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.665848] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 697.665848] env[61987]: value = "task-1061312" [ 697.665848] env[61987]: _type = "Task" [ 697.665848] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.669918] env[61987]: DEBUG nova.network.neutron [-] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.676238] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061312, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.688706] env[61987]: DEBUG nova.compute.manager [req-6f28123e-b6be-4eeb-897b-1fbb7293f4c4 req-9bdf44cc-1edd-44cc-ac1a-a58e8ec5e981 service nova] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Received event network-changed-a55b478b-de12-47cb-80e8-07a467463422 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 697.688936] env[61987]: DEBUG nova.compute.manager [req-6f28123e-b6be-4eeb-897b-1fbb7293f4c4 req-9bdf44cc-1edd-44cc-ac1a-a58e8ec5e981 service nova] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Refreshing instance network info cache due to event network-changed-a55b478b-de12-47cb-80e8-07a467463422. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 697.689194] env[61987]: DEBUG oslo_concurrency.lockutils [req-6f28123e-b6be-4eeb-897b-1fbb7293f4c4 req-9bdf44cc-1edd-44cc-ac1a-a58e8ec5e981 service nova] Acquiring lock "refresh_cache-02fe490b-75c8-406b-98e0-aa200f249673" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.689362] env[61987]: DEBUG oslo_concurrency.lockutils [req-6f28123e-b6be-4eeb-897b-1fbb7293f4c4 req-9bdf44cc-1edd-44cc-ac1a-a58e8ec5e981 service nova] Acquired lock "refresh_cache-02fe490b-75c8-406b-98e0-aa200f249673" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.689536] env[61987]: DEBUG nova.network.neutron [req-6f28123e-b6be-4eeb-897b-1fbb7293f4c4 req-9bdf44cc-1edd-44cc-ac1a-a58e8ec5e981 service nova] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Refreshing network info cache for port a55b478b-de12-47cb-80e8-07a467463422 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 697.866291] env[61987]: DEBUG nova.network.neutron [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Successfully updated port: 2b8e0044-50eb-40c8-8a96-9c828626d2f7 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 697.869920] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.620s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.870119] env[61987]: DEBUG nova.compute.manager [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 697.872587] env[61987]: DEBUG oslo_concurrency.lockutils [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.007s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.873981] env[61987]: INFO nova.compute.claims [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 698.024508] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061308, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.068213] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': task-1061309, 'name': PowerOnVM_Task} progress is 37%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.145516] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Creating Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 698.145998] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-5f947bbf-6367-4fc8-8f0a-ffb2fcd159b6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.154817] env[61987]: DEBUG oslo_vmware.api [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for the task: (returnval){ [ 698.154817] env[61987]: value = "task-1061313" [ 698.154817] env[61987]: _type = "Task" [ 698.154817] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.167162] env[61987]: DEBUG oslo_vmware.api [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061313, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.172555] env[61987]: INFO nova.compute.manager [-] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Took 1.51 seconds to deallocate network for instance. [ 698.181378] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061312, 'name': CreateVM_Task, 'duration_secs': 0.352949} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.185791] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 698.187071] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.187359] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.187822] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 698.188243] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b047fb3a-8d80-4fd4-9708-0be10845dc6f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.195303] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 698.195303] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52aa22c5-5159-1a28-7342-e4f574e55448" [ 698.195303] env[61987]: _type = "Task" [ 698.195303] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.208999] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52aa22c5-5159-1a28-7342-e4f574e55448, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.369618] env[61987]: DEBUG oslo_concurrency.lockutils [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Acquiring lock "refresh_cache-0b4b3bdb-7393-47ab-abc1-44c47496cb3a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.370113] env[61987]: DEBUG oslo_concurrency.lockutils [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Acquired lock "refresh_cache-0b4b3bdb-7393-47ab-abc1-44c47496cb3a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.370113] env[61987]: DEBUG nova.network.neutron [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 698.380134] env[61987]: DEBUG nova.compute.utils [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 698.382701] env[61987]: DEBUG nova.compute.manager [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 698.385747] env[61987]: DEBUG nova.network.neutron [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 698.479174] env[61987]: DEBUG nova.policy [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9073b17b601445fa9e4eb979debc2f4d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3fd98395e57149c987f82585296751b2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 698.526423] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061308, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.921235} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.528920] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Copied Virtual Disk [datastore1] vmware_temp/e3eaa15d-e0db-416b-af95-91e76c2b660e/7bf8969f-d179-42af-9e15-527d5449f929/tmp-sparse.vmdk to [datastore1] vmware_temp/e3eaa15d-e0db-416b-af95-91e76c2b660e/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 698.529109] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Deleting the datastore file [datastore1] vmware_temp/e3eaa15d-e0db-416b-af95-91e76c2b660e/7bf8969f-d179-42af-9e15-527d5449f929/tmp-sparse.vmdk {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 698.529373] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-351ab9ab-d257-466f-9be9-59e6ce4d2e04 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.537543] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 698.537543] env[61987]: value = "task-1061314" [ 698.537543] env[61987]: _type = "Task" [ 698.537543] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.546354] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061314, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.568632] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': task-1061309, 'name': PowerOnVM_Task} progress is 82%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.596374] env[61987]: DEBUG nova.network.neutron [req-6f28123e-b6be-4eeb-897b-1fbb7293f4c4 req-9bdf44cc-1edd-44cc-ac1a-a58e8ec5e981 service nova] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Updated VIF entry in instance network info cache for port a55b478b-de12-47cb-80e8-07a467463422. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 698.596949] env[61987]: DEBUG nova.network.neutron [req-6f28123e-b6be-4eeb-897b-1fbb7293f4c4 req-9bdf44cc-1edd-44cc-ac1a-a58e8ec5e981 service nova] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Updating instance_info_cache with network_info: [{"id": "a55b478b-de12-47cb-80e8-07a467463422", "address": "fa:16:3e:b5:73:24", "network": {"id": "e2d0ce90-ff14-47de-bc4e-a2564d1268a7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-222691221-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d828262de7b14d2ab38193d7d34e8f7a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa55b478b-de", "ovs_interfaceid": "a55b478b-de12-47cb-80e8-07a467463422", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.670037] env[61987]: DEBUG oslo_vmware.api [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061313, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.688802] env[61987]: DEBUG oslo_concurrency.lockutils [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.711034] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.711034] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 698.711807] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.893542] env[61987]: DEBUG nova.compute.manager [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 698.900700] env[61987]: DEBUG nova.network.neutron [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Successfully created port: 4e18ece6-5812-4c5c-827d-b3ee7d399008 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 698.938282] env[61987]: DEBUG nova.network.neutron [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 699.047594] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061314, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.047724} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.048291] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 699.048291] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Moving file from [datastore1] vmware_temp/e3eaa15d-e0db-416b-af95-91e76c2b660e/7bf8969f-d179-42af-9e15-527d5449f929 to [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929. {{(pid=61987) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 699.048492] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-8931521b-944d-41bd-8d3c-af90ca5903d0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.055842] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 699.055842] env[61987]: value = "task-1061315" [ 699.055842] env[61987]: _type = "Task" [ 699.055842] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.067610] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061315, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.075727] env[61987]: DEBUG oslo_vmware.api [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': task-1061309, 'name': PowerOnVM_Task, 'duration_secs': 1.162643} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.076452] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 699.076452] env[61987]: INFO nova.compute.manager [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Took 9.21 seconds to spawn the instance on the hypervisor. [ 699.076661] env[61987]: DEBUG nova.compute.manager [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 699.077658] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f36baa7-0889-43e4-92b3-b163ac42b8e5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.106025] env[61987]: DEBUG oslo_concurrency.lockutils [req-6f28123e-b6be-4eeb-897b-1fbb7293f4c4 req-9bdf44cc-1edd-44cc-ac1a-a58e8ec5e981 service nova] Releasing lock "refresh_cache-02fe490b-75c8-406b-98e0-aa200f249673" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.106025] env[61987]: DEBUG nova.compute.manager [req-6f28123e-b6be-4eeb-897b-1fbb7293f4c4 req-9bdf44cc-1edd-44cc-ac1a-a58e8ec5e981 service nova] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Received event network-vif-deleted-cf9c4ef3-c7ef-4194-bd5b-72f4bafee5fa {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 699.106025] env[61987]: INFO nova.compute.manager [req-6f28123e-b6be-4eeb-897b-1fbb7293f4c4 req-9bdf44cc-1edd-44cc-ac1a-a58e8ec5e981 service nova] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Neutron deleted interface cf9c4ef3-c7ef-4194-bd5b-72f4bafee5fa; detaching it from the instance and deleting it from the info cache [ 699.106025] env[61987]: DEBUG nova.network.neutron [req-6f28123e-b6be-4eeb-897b-1fbb7293f4c4 req-9bdf44cc-1edd-44cc-ac1a-a58e8ec5e981 service nova] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.172352] env[61987]: DEBUG oslo_vmware.api [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061313, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.287353] env[61987]: DEBUG nova.network.neutron [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Updating instance_info_cache with network_info: [{"id": "2b8e0044-50eb-40c8-8a96-9c828626d2f7", "address": "fa:16:3e:35:00:90", "network": {"id": "cb282746-db11-437f-a604-71af177feab6", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1203425515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c062bbfea30a42bbb2dbec5a84358d95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b8e0044-50", "ovs_interfaceid": "2b8e0044-50eb-40c8-8a96-9c828626d2f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.351130] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94040697-fb55-4e00-a870-82251f5d3b88 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.359612] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e78f22e3-4f2c-4267-ac81-9ffce783f026 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.405406] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d985688c-6b03-4b09-be90-e3677ab59841 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.416231] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a41d6ca-a788-4429-80c5-101466bc4936 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.438023] env[61987]: DEBUG nova.compute.provider_tree [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 699.565926] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061315, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.057719} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.566214] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] File moved {{(pid=61987) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 699.566417] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Cleaning up location [datastore1] vmware_temp/e3eaa15d-e0db-416b-af95-91e76c2b660e {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 699.566578] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Deleting the datastore file [datastore1] vmware_temp/e3eaa15d-e0db-416b-af95-91e76c2b660e {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 699.566829] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-58c21e8f-6e64-44b4-9d02-b4747453f1ce {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.573533] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 699.573533] env[61987]: value = "task-1061316" [ 699.573533] env[61987]: _type = "Task" [ 699.573533] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.581626] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061316, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.598961] env[61987]: INFO nova.compute.manager [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Took 42.11 seconds to build instance. [ 699.606720] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8c50b2f1-997f-477a-9b12-c01b0b6953f5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.615069] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcf8adf5-2d0e-4296-b559-27bcfb1ea9fa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.641205] env[61987]: DEBUG nova.compute.manager [req-6f28123e-b6be-4eeb-897b-1fbb7293f4c4 req-9bdf44cc-1edd-44cc-ac1a-a58e8ec5e981 service nova] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Detach interface failed, port_id=cf9c4ef3-c7ef-4194-bd5b-72f4bafee5fa, reason: Instance 89a97f81-bcd9-49e5-bbd3-9140f6638a20 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 699.664953] env[61987]: DEBUG oslo_vmware.api [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061313, 'name': CreateSnapshot_Task, 'duration_secs': 1.504419} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.665103] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Created Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 699.665875] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f9a65c5-9f78-4767-9ecd-d601041081eb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.767111] env[61987]: DEBUG nova.compute.manager [req-803b7ea4-90da-489c-bd08-adbb7c104717 req-7c72924d-05b3-4efc-9713-5c7f3bbd9bb7 service nova] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Received event network-vif-plugged-2b8e0044-50eb-40c8-8a96-9c828626d2f7 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 699.767340] env[61987]: DEBUG oslo_concurrency.lockutils [req-803b7ea4-90da-489c-bd08-adbb7c104717 req-7c72924d-05b3-4efc-9713-5c7f3bbd9bb7 service nova] Acquiring lock "0b4b3bdb-7393-47ab-abc1-44c47496cb3a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.767543] env[61987]: DEBUG oslo_concurrency.lockutils [req-803b7ea4-90da-489c-bd08-adbb7c104717 req-7c72924d-05b3-4efc-9713-5c7f3bbd9bb7 service nova] Lock "0b4b3bdb-7393-47ab-abc1-44c47496cb3a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.767709] env[61987]: DEBUG oslo_concurrency.lockutils [req-803b7ea4-90da-489c-bd08-adbb7c104717 req-7c72924d-05b3-4efc-9713-5c7f3bbd9bb7 service nova] Lock "0b4b3bdb-7393-47ab-abc1-44c47496cb3a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.767881] env[61987]: DEBUG nova.compute.manager [req-803b7ea4-90da-489c-bd08-adbb7c104717 req-7c72924d-05b3-4efc-9713-5c7f3bbd9bb7 service nova] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] No waiting events found dispatching network-vif-plugged-2b8e0044-50eb-40c8-8a96-9c828626d2f7 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 699.768366] env[61987]: WARNING nova.compute.manager [req-803b7ea4-90da-489c-bd08-adbb7c104717 req-7c72924d-05b3-4efc-9713-5c7f3bbd9bb7 service nova] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Received unexpected event network-vif-plugged-2b8e0044-50eb-40c8-8a96-9c828626d2f7 for instance with vm_state building and task_state spawning. [ 699.768628] env[61987]: DEBUG nova.compute.manager [req-803b7ea4-90da-489c-bd08-adbb7c104717 req-7c72924d-05b3-4efc-9713-5c7f3bbd9bb7 service nova] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Received event network-changed-2b8e0044-50eb-40c8-8a96-9c828626d2f7 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 699.768773] env[61987]: DEBUG nova.compute.manager [req-803b7ea4-90da-489c-bd08-adbb7c104717 req-7c72924d-05b3-4efc-9713-5c7f3bbd9bb7 service nova] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Refreshing instance network info cache due to event network-changed-2b8e0044-50eb-40c8-8a96-9c828626d2f7. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 699.768954] env[61987]: DEBUG oslo_concurrency.lockutils [req-803b7ea4-90da-489c-bd08-adbb7c104717 req-7c72924d-05b3-4efc-9713-5c7f3bbd9bb7 service nova] Acquiring lock "refresh_cache-0b4b3bdb-7393-47ab-abc1-44c47496cb3a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.789581] env[61987]: DEBUG oslo_concurrency.lockutils [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Releasing lock "refresh_cache-0b4b3bdb-7393-47ab-abc1-44c47496cb3a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.789911] env[61987]: DEBUG nova.compute.manager [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Instance network_info: |[{"id": "2b8e0044-50eb-40c8-8a96-9c828626d2f7", "address": "fa:16:3e:35:00:90", "network": {"id": "cb282746-db11-437f-a604-71af177feab6", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1203425515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c062bbfea30a42bbb2dbec5a84358d95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b8e0044-50", "ovs_interfaceid": "2b8e0044-50eb-40c8-8a96-9c828626d2f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 699.790303] env[61987]: DEBUG oslo_concurrency.lockutils [req-803b7ea4-90da-489c-bd08-adbb7c104717 req-7c72924d-05b3-4efc-9713-5c7f3bbd9bb7 service nova] Acquired lock "refresh_cache-0b4b3bdb-7393-47ab-abc1-44c47496cb3a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.790453] env[61987]: DEBUG nova.network.neutron [req-803b7ea4-90da-489c-bd08-adbb7c104717 req-7c72924d-05b3-4efc-9713-5c7f3bbd9bb7 service nova] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Refreshing network info cache for port 2b8e0044-50eb-40c8-8a96-9c828626d2f7 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 699.791943] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:35:00:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3a0ddd7d-c321-4187-bdd8-b19044ea2c4a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2b8e0044-50eb-40c8-8a96-9c828626d2f7', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 699.799721] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Creating folder: Project (c062bbfea30a42bbb2dbec5a84358d95). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 699.804756] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b212d85b-41fe-4b76-ae3d-df3881ada627 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.814850] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Created folder: Project (c062bbfea30a42bbb2dbec5a84358d95) in parent group-v234219. [ 699.814850] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Creating folder: Instances. Parent ref: group-v234249. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 699.814973] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6eca529d-c613-4cf1-a3a0-793eca39d9d0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.826575] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Created folder: Instances in parent group-v234249. [ 699.826721] env[61987]: DEBUG oslo.service.loopingcall [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 699.826914] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 699.827206] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b4fcfc41-b4cf-41a6-9ec6-2c897ab4a66b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.850118] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 699.850118] env[61987]: value = "task-1061319" [ 699.850118] env[61987]: _type = "Task" [ 699.850118] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.857536] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061319, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.905996] env[61987]: DEBUG nova.compute.manager [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 699.934587] env[61987]: DEBUG nova.virt.hardware [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 699.934873] env[61987]: DEBUG nova.virt.hardware [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 699.935067] env[61987]: DEBUG nova.virt.hardware [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 699.935246] env[61987]: DEBUG nova.virt.hardware [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 699.935403] env[61987]: DEBUG nova.virt.hardware [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 699.936263] env[61987]: DEBUG nova.virt.hardware [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 699.936263] env[61987]: DEBUG nova.virt.hardware [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 699.936263] env[61987]: DEBUG nova.virt.hardware [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 699.936263] env[61987]: DEBUG nova.virt.hardware [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 699.936263] env[61987]: DEBUG nova.virt.hardware [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 699.936465] env[61987]: DEBUG nova.virt.hardware [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 699.937395] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccf2b481-0752-4ba6-bbed-0a5bad241389 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.940922] env[61987]: DEBUG nova.scheduler.client.report [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 699.949709] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88e96eb3-e047-4387-bd22-9dc0bada5bbd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.083797] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061316, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.029107} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.084075] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 700.084852] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-650ae428-0600-4e08-9c91-36896f38f202 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.090123] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 700.090123] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52018a4d-e7d9-0fc7-5c20-d7594e7e3fa2" [ 700.090123] env[61987]: _type = "Task" [ 700.090123] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.098525] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52018a4d-e7d9-0fc7-5c20-d7594e7e3fa2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.101085] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3e44722-18ae-48c9-a727-ddb42e822bf0 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Lock "466a5f5a-f48a-4921-8bcb-44a1840c8141" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.406s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.107880] env[61987]: DEBUG nova.network.neutron [req-803b7ea4-90da-489c-bd08-adbb7c104717 req-7c72924d-05b3-4efc-9713-5c7f3bbd9bb7 service nova] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Updated VIF entry in instance network info cache for port 2b8e0044-50eb-40c8-8a96-9c828626d2f7. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 700.108221] env[61987]: DEBUG nova.network.neutron [req-803b7ea4-90da-489c-bd08-adbb7c104717 req-7c72924d-05b3-4efc-9713-5c7f3bbd9bb7 service nova] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Updating instance_info_cache with network_info: [{"id": "2b8e0044-50eb-40c8-8a96-9c828626d2f7", "address": "fa:16:3e:35:00:90", "network": {"id": "cb282746-db11-437f-a604-71af177feab6", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1203425515-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c062bbfea30a42bbb2dbec5a84358d95", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3a0ddd7d-c321-4187-bdd8-b19044ea2c4a", "external-id": "nsx-vlan-transportzone-747", "segmentation_id": 747, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b8e0044-50", "ovs_interfaceid": "2b8e0044-50eb-40c8-8a96-9c828626d2f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.190406] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Creating linked-clone VM from snapshot {{(pid=61987) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 700.191198] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0153d8c1-6b23-42be-985c-97b8de9d5a15 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.200507] env[61987]: DEBUG oslo_vmware.api [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for the task: (returnval){ [ 700.200507] env[61987]: value = "task-1061320" [ 700.200507] env[61987]: _type = "Task" [ 700.200507] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.208723] env[61987]: DEBUG oslo_vmware.api [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061320, 'name': CloneVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.229008] env[61987]: DEBUG nova.compute.manager [req-b759628f-ca6d-4626-be0c-866e081168b7 req-e4ecc060-f5a3-48d1-a9d1-16c91ffa504d service nova] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Received event network-changed-39a5fbe3-c262-40f4-acdc-48209a7f617a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 700.229392] env[61987]: DEBUG nova.compute.manager [req-b759628f-ca6d-4626-be0c-866e081168b7 req-e4ecc060-f5a3-48d1-a9d1-16c91ffa504d service nova] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Refreshing instance network info cache due to event network-changed-39a5fbe3-c262-40f4-acdc-48209a7f617a. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 700.229567] env[61987]: DEBUG oslo_concurrency.lockutils [req-b759628f-ca6d-4626-be0c-866e081168b7 req-e4ecc060-f5a3-48d1-a9d1-16c91ffa504d service nova] Acquiring lock "refresh_cache-466a5f5a-f48a-4921-8bcb-44a1840c8141" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.229739] env[61987]: DEBUG oslo_concurrency.lockutils [req-b759628f-ca6d-4626-be0c-866e081168b7 req-e4ecc060-f5a3-48d1-a9d1-16c91ffa504d service nova] Acquired lock "refresh_cache-466a5f5a-f48a-4921-8bcb-44a1840c8141" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.229912] env[61987]: DEBUG nova.network.neutron [req-b759628f-ca6d-4626-be0c-866e081168b7 req-e4ecc060-f5a3-48d1-a9d1-16c91ffa504d service nova] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Refreshing network info cache for port 39a5fbe3-c262-40f4-acdc-48209a7f617a {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 700.361419] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061319, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.447025] env[61987]: DEBUG oslo_concurrency.lockutils [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.574s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.447025] env[61987]: DEBUG nova.compute.manager [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 700.453035] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.607s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.454581] env[61987]: INFO nova.compute.claims [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 700.604018] env[61987]: DEBUG nova.compute.manager [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 700.607518] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52018a4d-e7d9-0fc7-5c20-d7594e7e3fa2, 'name': SearchDatastore_Task, 'duration_secs': 0.010438} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.608843] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.609189] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] ea6c2120-a69a-42c7-bc92-3e74539c008d/ea6c2120-a69a-42c7-bc92-3e74539c008d.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 700.609518] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.610430] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 700.610723] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c7785d6c-4cb6-4b83-916f-f253b3b14ebd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.614205] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c7b739fd-f610-4168-83f4-c615fd1fdf41 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.616512] env[61987]: DEBUG oslo_concurrency.lockutils [req-803b7ea4-90da-489c-bd08-adbb7c104717 req-7c72924d-05b3-4efc-9713-5c7f3bbd9bb7 service nova] Releasing lock "refresh_cache-0b4b3bdb-7393-47ab-abc1-44c47496cb3a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.624810] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 700.624810] env[61987]: value = "task-1061321" [ 700.624810] env[61987]: _type = "Task" [ 700.624810] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.629024] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 700.629024] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 700.629024] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e9cb853-67e6-45cd-9a63-69b5433df892 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.633877] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061321, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.637780] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 700.637780] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]524acf53-28ac-1e12-9d12-4f63a4f3b96e" [ 700.637780] env[61987]: _type = "Task" [ 700.637780] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.645606] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524acf53-28ac-1e12-9d12-4f63a4f3b96e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.712555] env[61987]: DEBUG oslo_vmware.api [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061320, 'name': CloneVM_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.735439] env[61987]: DEBUG nova.network.neutron [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Successfully updated port: 4e18ece6-5812-4c5c-827d-b3ee7d399008 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 700.864281] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061319, 'name': CreateVM_Task, 'duration_secs': 0.72762} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.864509] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 700.865188] env[61987]: DEBUG oslo_concurrency.lockutils [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.865359] env[61987]: DEBUG oslo_concurrency.lockutils [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.865699] env[61987]: DEBUG oslo_concurrency.lockutils [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 700.865968] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb4f2ec8-4d91-444c-997e-391bd7cd7fce {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.870947] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Waiting for the task: (returnval){ [ 700.870947] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52c633d4-94fe-ccf1-d8df-9a087fbef174" [ 700.870947] env[61987]: _type = "Task" [ 700.870947] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.881564] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c633d4-94fe-ccf1-d8df-9a087fbef174, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.956033] env[61987]: DEBUG nova.compute.utils [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 700.956033] env[61987]: DEBUG nova.compute.manager [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 700.956588] env[61987]: DEBUG nova.network.neutron [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 701.134262] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061321, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.466478} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.134828] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] ea6c2120-a69a-42c7-bc92-3e74539c008d/ea6c2120-a69a-42c7-bc92-3e74539c008d.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 701.135335] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 701.135843] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d3ef796e-e999-41bc-b7f9-62e2cd8a48a0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.150131] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524acf53-28ac-1e12-9d12-4f63a4f3b96e, 'name': SearchDatastore_Task, 'duration_secs': 0.008719} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.152175] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 701.152175] env[61987]: value = "task-1061322" [ 701.152175] env[61987]: _type = "Task" [ 701.152175] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.152703] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e400cba4-e4ec-49e1-a6c7-573ee91b8f3e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.165202] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.166414] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 701.166414] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52371959-cb43-f176-ee1d-57fe054c00a3" [ 701.166414] env[61987]: _type = "Task" [ 701.166414] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.172323] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061322, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.175115] env[61987]: DEBUG nova.policy [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '555cc0efbb98415f8f5f2277695b5484', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bbd655fdd1f54787bf2056369899c924', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 701.184772] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52371959-cb43-f176-ee1d-57fe054c00a3, 'name': SearchDatastore_Task, 'duration_secs': 0.01014} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.185219] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.185634] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 02fe490b-75c8-406b-98e0-aa200f249673/02fe490b-75c8-406b-98e0-aa200f249673.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 701.188025] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-90de69c1-97e7-4934-b12a-4a4816d6597e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.191987] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 701.191987] env[61987]: value = "task-1061323" [ 701.191987] env[61987]: _type = "Task" [ 701.191987] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.203358] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061323, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.213824] env[61987]: DEBUG oslo_vmware.api [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061320, 'name': CloneVM_Task} progress is 95%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.239465] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "refresh_cache-51a36102-795e-47b7-b96a-857e54dc703e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.239465] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquired lock "refresh_cache-51a36102-795e-47b7-b96a-857e54dc703e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.239465] env[61987]: DEBUG nova.network.neutron [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 701.382681] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c633d4-94fe-ccf1-d8df-9a087fbef174, 'name': SearchDatastore_Task, 'duration_secs': 0.07577} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.383778] env[61987]: DEBUG nova.network.neutron [req-b759628f-ca6d-4626-be0c-866e081168b7 req-e4ecc060-f5a3-48d1-a9d1-16c91ffa504d service nova] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Updated VIF entry in instance network info cache for port 39a5fbe3-c262-40f4-acdc-48209a7f617a. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 701.385451] env[61987]: DEBUG nova.network.neutron [req-b759628f-ca6d-4626-be0c-866e081168b7 req-e4ecc060-f5a3-48d1-a9d1-16c91ffa504d service nova] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Updating instance_info_cache with network_info: [{"id": "39a5fbe3-c262-40f4-acdc-48209a7f617a", "address": "fa:16:3e:f9:43:9a", "network": {"id": "020556c5-6e13-45fa-a715-bf097bdf0f73", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-367382123-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3adce5cde814bf0bb57b8e941dba79f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56834f67-27a8-43dc-bbc6-a74aaa08959b", "external-id": "nsx-vlan-transportzone-949", "segmentation_id": 949, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39a5fbe3-c2", "ovs_interfaceid": "39a5fbe3-c262-40f4-acdc-48209a7f617a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.386199] env[61987]: DEBUG oslo_concurrency.lockutils [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.386583] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 701.387106] env[61987]: DEBUG oslo_concurrency.lockutils [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.387413] env[61987]: DEBUG oslo_concurrency.lockutils [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.389707] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 701.389707] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7f2dbe90-ba42-4334-9705-799cf27f24d1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.397518] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 701.398033] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 701.400019] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c935655-438e-4f3a-be33-e7d2c886df3e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.405242] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Waiting for the task: (returnval){ [ 701.405242] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52896eb2-32c6-09fe-6a65-e400e8d7fa74" [ 701.405242] env[61987]: _type = "Task" [ 701.405242] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.413544] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52896eb2-32c6-09fe-6a65-e400e8d7fa74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.460454] env[61987]: DEBUG nova.compute.manager [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 701.668505] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061322, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061364} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.668775] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 701.669613] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12cb2c9b-05fc-45df-b39e-5ff2b56aad81 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.693668] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Reconfiguring VM instance instance-00000023 to attach disk [datastore1] ea6c2120-a69a-42c7-bc92-3e74539c008d/ea6c2120-a69a-42c7-bc92-3e74539c008d.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 701.697276] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d71242f6-4c81-4842-a072-535381019caf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.737025] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 701.737025] env[61987]: value = "task-1061324" [ 701.737025] env[61987]: _type = "Task" [ 701.737025] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.737025] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061323, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.737025] env[61987]: DEBUG oslo_vmware.api [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061320, 'name': CloneVM_Task, 'duration_secs': 1.451616} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.737025] env[61987]: INFO nova.virt.vmwareapi.vmops [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Created linked-clone VM from snapshot [ 701.739288] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-294a66a3-b7ed-43ce-9e73-c85be6c9289f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.756199] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061324, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.756835] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Uploading image d25beb0f-8ad6-4e15-bf01-99793b52926f {{(pid=61987) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 701.785477] env[61987]: DEBUG oslo_vmware.rw_handles [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 701.785477] env[61987]: value = "vm-234252" [ 701.785477] env[61987]: _type = "VirtualMachine" [ 701.785477] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 701.786245] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-64b65176-2062-4b36-8408-c1ecbe5516e2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.797352] env[61987]: DEBUG oslo_vmware.rw_handles [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Lease: (returnval){ [ 701.797352] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52d60139-1d6c-c301-d60a-d5b170259faa" [ 701.797352] env[61987]: _type = "HttpNfcLease" [ 701.797352] env[61987]: } obtained for exporting VM: (result){ [ 701.797352] env[61987]: value = "vm-234252" [ 701.797352] env[61987]: _type = "VirtualMachine" [ 701.797352] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 701.797730] env[61987]: DEBUG oslo_vmware.api [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for the lease: (returnval){ [ 701.797730] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52d60139-1d6c-c301-d60a-d5b170259faa" [ 701.797730] env[61987]: _type = "HttpNfcLease" [ 701.797730] env[61987]: } to be ready. {{(pid=61987) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 701.807587] env[61987]: DEBUG nova.network.neutron [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 701.810542] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 701.810542] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52d60139-1d6c-c301-d60a-d5b170259faa" [ 701.810542] env[61987]: _type = "HttpNfcLease" [ 701.810542] env[61987]: } is initializing. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 701.889296] env[61987]: DEBUG oslo_concurrency.lockutils [req-b759628f-ca6d-4626-be0c-866e081168b7 req-e4ecc060-f5a3-48d1-a9d1-16c91ffa504d service nova] Releasing lock "refresh_cache-466a5f5a-f48a-4921-8bcb-44a1840c8141" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.911750] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b91b475-4b18-4551-bac9-9eae81c4c54f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.924746] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12cbfb32-ba84-43aa-8308-e6b60c5b0698 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.927517] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52896eb2-32c6-09fe-6a65-e400e8d7fa74, 'name': SearchDatastore_Task, 'duration_secs': 0.014843} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.928937] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb45aadc-1cf7-4c8a-b697-0ded519b37cf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.960200] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-783bfe55-a16b-4bc8-9564-abe0ef4f4289 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.965613] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Waiting for the task: (returnval){ [ 701.965613] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]522c1eed-e78f-d849-22c6-701854978696" [ 701.965613] env[61987]: _type = "Task" [ 701.965613] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.976194] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d507c8d-4b3b-44d6-bc54-92ff6950ffae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.981670] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]522c1eed-e78f-d849-22c6-701854978696, 'name': SearchDatastore_Task, 'duration_secs': 0.023181} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.982304] env[61987]: DEBUG oslo_concurrency.lockutils [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.982594] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 0b4b3bdb-7393-47ab-abc1-44c47496cb3a/0b4b3bdb-7393-47ab-abc1-44c47496cb3a.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 701.982894] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-da54baf2-7724-4d02-be26-f1e6e90a7397 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.995449] env[61987]: DEBUG nova.compute.provider_tree [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.000396] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Waiting for the task: (returnval){ [ 702.000396] env[61987]: value = "task-1061326" [ 702.000396] env[61987]: _type = "Task" [ 702.000396] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.008553] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': task-1061326, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.098811] env[61987]: DEBUG nova.compute.manager [req-537d414c-1f2e-4c7f-b047-5ce8e1d47700 req-8ff89854-366c-469a-a8d0-ddce439ad371 service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Received event network-vif-plugged-4e18ece6-5812-4c5c-827d-b3ee7d399008 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 702.099045] env[61987]: DEBUG oslo_concurrency.lockutils [req-537d414c-1f2e-4c7f-b047-5ce8e1d47700 req-8ff89854-366c-469a-a8d0-ddce439ad371 service nova] Acquiring lock "51a36102-795e-47b7-b96a-857e54dc703e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.099258] env[61987]: DEBUG oslo_concurrency.lockutils [req-537d414c-1f2e-4c7f-b047-5ce8e1d47700 req-8ff89854-366c-469a-a8d0-ddce439ad371 service nova] Lock "51a36102-795e-47b7-b96a-857e54dc703e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.099454] env[61987]: DEBUG oslo_concurrency.lockutils [req-537d414c-1f2e-4c7f-b047-5ce8e1d47700 req-8ff89854-366c-469a-a8d0-ddce439ad371 service nova] Lock "51a36102-795e-47b7-b96a-857e54dc703e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.099589] env[61987]: DEBUG nova.compute.manager [req-537d414c-1f2e-4c7f-b047-5ce8e1d47700 req-8ff89854-366c-469a-a8d0-ddce439ad371 service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] No waiting events found dispatching network-vif-plugged-4e18ece6-5812-4c5c-827d-b3ee7d399008 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 702.099771] env[61987]: WARNING nova.compute.manager [req-537d414c-1f2e-4c7f-b047-5ce8e1d47700 req-8ff89854-366c-469a-a8d0-ddce439ad371 service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Received unexpected event network-vif-plugged-4e18ece6-5812-4c5c-827d-b3ee7d399008 for instance with vm_state building and task_state spawning. [ 702.099943] env[61987]: DEBUG nova.compute.manager [req-537d414c-1f2e-4c7f-b047-5ce8e1d47700 req-8ff89854-366c-469a-a8d0-ddce439ad371 service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Received event network-changed-4e18ece6-5812-4c5c-827d-b3ee7d399008 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 702.100116] env[61987]: DEBUG nova.compute.manager [req-537d414c-1f2e-4c7f-b047-5ce8e1d47700 req-8ff89854-366c-469a-a8d0-ddce439ad371 service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Refreshing instance network info cache due to event network-changed-4e18ece6-5812-4c5c-827d-b3ee7d399008. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 702.100284] env[61987]: DEBUG oslo_concurrency.lockutils [req-537d414c-1f2e-4c7f-b047-5ce8e1d47700 req-8ff89854-366c-469a-a8d0-ddce439ad371 service nova] Acquiring lock "refresh_cache-51a36102-795e-47b7-b96a-857e54dc703e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.134019] env[61987]: DEBUG nova.network.neutron [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Updating instance_info_cache with network_info: [{"id": "4e18ece6-5812-4c5c-827d-b3ee7d399008", "address": "fa:16:3e:1a:e6:c9", "network": {"id": "abcd2857-4c5a-4546-a003-858ddb17b0e6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-124838733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3fd98395e57149c987f82585296751b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e18ece6-58", "ovs_interfaceid": "4e18ece6-5812-4c5c-827d-b3ee7d399008", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.210287] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061323, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.622648} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.210579] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 02fe490b-75c8-406b-98e0-aa200f249673/02fe490b-75c8-406b-98e0-aa200f249673.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 702.210823] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 702.211087] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2e9fe5a4-fe2e-463e-bb47-4dba6a9f90a3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.217887] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 702.217887] env[61987]: value = "task-1061327" [ 702.217887] env[61987]: _type = "Task" [ 702.217887] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.223502] env[61987]: DEBUG nova.network.neutron [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Successfully created port: 5134697b-ec80-40fc-9a81-5f002c9421bc {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 702.229498] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061327, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.243019] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061324, 'name': ReconfigVM_Task, 'duration_secs': 0.340542} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.243273] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Reconfigured VM instance instance-00000023 to attach disk [datastore1] ea6c2120-a69a-42c7-bc92-3e74539c008d/ea6c2120-a69a-42c7-bc92-3e74539c008d.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 702.244078] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa6573f2-c4f5-4ef8-bcd0-4cb018a25ced {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.251686] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 702.251686] env[61987]: value = "task-1061328" [ 702.251686] env[61987]: _type = "Task" [ 702.251686] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.258124] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061328, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.306096] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 702.306096] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52d60139-1d6c-c301-d60a-d5b170259faa" [ 702.306096] env[61987]: _type = "HttpNfcLease" [ 702.306096] env[61987]: } is ready. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 702.306502] env[61987]: DEBUG oslo_vmware.rw_handles [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 702.306502] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52d60139-1d6c-c301-d60a-d5b170259faa" [ 702.306502] env[61987]: _type = "HttpNfcLease" [ 702.306502] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 702.307628] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f15e0b8-f380-4e15-be36-2f0150ebf713 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.315140] env[61987]: DEBUG oslo_vmware.rw_handles [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52984d26-9de9-3da9-53cb-5a3fdc3d9cc4/disk-0.vmdk from lease info. {{(pid=61987) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 702.315402] env[61987]: DEBUG oslo_vmware.rw_handles [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52984d26-9de9-3da9-53cb-5a3fdc3d9cc4/disk-0.vmdk for reading. {{(pid=61987) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 702.475577] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2c9c10b1-1f62-41e8-9dbc-c8b36039acd5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.479470] env[61987]: DEBUG nova.compute.manager [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 702.507441] env[61987]: DEBUG nova.scheduler.client.report [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 702.523399] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': task-1061326, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.527880] env[61987]: DEBUG nova.virt.hardware [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 702.528155] env[61987]: DEBUG nova.virt.hardware [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 702.528318] env[61987]: DEBUG nova.virt.hardware [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 702.528503] env[61987]: DEBUG nova.virt.hardware [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 702.528653] env[61987]: DEBUG nova.virt.hardware [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 702.528804] env[61987]: DEBUG nova.virt.hardware [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 702.529016] env[61987]: DEBUG nova.virt.hardware [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 702.529191] env[61987]: DEBUG nova.virt.hardware [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 702.529358] env[61987]: DEBUG nova.virt.hardware [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 702.530149] env[61987]: DEBUG nova.virt.hardware [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 702.530149] env[61987]: DEBUG nova.virt.hardware [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 702.530609] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d92bb53c-4491-4293-bef7-457b5a258969 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.543597] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29c1feab-c710-4c5c-bbec-4d088d780326 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.636240] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Releasing lock "refresh_cache-51a36102-795e-47b7-b96a-857e54dc703e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.636589] env[61987]: DEBUG nova.compute.manager [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Instance network_info: |[{"id": "4e18ece6-5812-4c5c-827d-b3ee7d399008", "address": "fa:16:3e:1a:e6:c9", "network": {"id": "abcd2857-4c5a-4546-a003-858ddb17b0e6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-124838733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3fd98395e57149c987f82585296751b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e18ece6-58", "ovs_interfaceid": "4e18ece6-5812-4c5c-827d-b3ee7d399008", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 702.636923] env[61987]: DEBUG oslo_concurrency.lockutils [req-537d414c-1f2e-4c7f-b047-5ce8e1d47700 req-8ff89854-366c-469a-a8d0-ddce439ad371 service nova] Acquired lock "refresh_cache-51a36102-795e-47b7-b96a-857e54dc703e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.637115] env[61987]: DEBUG nova.network.neutron [req-537d414c-1f2e-4c7f-b047-5ce8e1d47700 req-8ff89854-366c-469a-a8d0-ddce439ad371 service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Refreshing network info cache for port 4e18ece6-5812-4c5c-827d-b3ee7d399008 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 702.638386] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1a:e6:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '170f3b82-5915-4e36-bce9-4664ebb6be5e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4e18ece6-5812-4c5c-827d-b3ee7d399008', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 702.645935] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Creating folder: Project (3fd98395e57149c987f82585296751b2). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 702.649244] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-195489be-f5df-49a3-981a-b0fbdbff1aca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.663174] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Created folder: Project (3fd98395e57149c987f82585296751b2) in parent group-v234219. [ 702.663360] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Creating folder: Instances. Parent ref: group-v234253. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 702.663649] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f4f455b2-d2be-470e-bd55-c5e286dd15ce {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.674515] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Created folder: Instances in parent group-v234253. [ 702.674807] env[61987]: DEBUG oslo.service.loopingcall [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 702.675176] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 702.675255] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-941cf2d3-0637-41cc-af76-0054372b427e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.697936] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 702.697936] env[61987]: value = "task-1061331" [ 702.697936] env[61987]: _type = "Task" [ 702.697936] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.707780] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061331, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.728306] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061327, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0652} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.729048] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 702.729643] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03d1b333-e9c0-4023-8f48-20b2314061d0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.755365] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Reconfiguring VM instance instance-00000024 to attach disk [datastore1] 02fe490b-75c8-406b-98e0-aa200f249673/02fe490b-75c8-406b-98e0-aa200f249673.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 702.758101] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cab37ef7-cdfd-4d44-b568-0edaaf6cd142 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.781582] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061328, 'name': Rename_Task, 'duration_secs': 0.131181} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.781882] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 702.782163] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e5a5809b-3dca-426f-8461-b2ec530eabc0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.785044] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 702.785044] env[61987]: value = "task-1061332" [ 702.785044] env[61987]: _type = "Task" [ 702.785044] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.789525] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 702.789525] env[61987]: value = "task-1061333" [ 702.789525] env[61987]: _type = "Task" [ 702.789525] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.796162] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061332, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.801803] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061333, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.906850] env[61987]: DEBUG nova.network.neutron [req-537d414c-1f2e-4c7f-b047-5ce8e1d47700 req-8ff89854-366c-469a-a8d0-ddce439ad371 service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Updated VIF entry in instance network info cache for port 4e18ece6-5812-4c5c-827d-b3ee7d399008. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 702.907389] env[61987]: DEBUG nova.network.neutron [req-537d414c-1f2e-4c7f-b047-5ce8e1d47700 req-8ff89854-366c-469a-a8d0-ddce439ad371 service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Updating instance_info_cache with network_info: [{"id": "4e18ece6-5812-4c5c-827d-b3ee7d399008", "address": "fa:16:3e:1a:e6:c9", "network": {"id": "abcd2857-4c5a-4546-a003-858ddb17b0e6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-124838733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3fd98395e57149c987f82585296751b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e18ece6-58", "ovs_interfaceid": "4e18ece6-5812-4c5c-827d-b3ee7d399008", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.022410] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.569s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.024014] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': task-1061326, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.024671] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.836s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.026166] env[61987]: INFO nova.compute.claims [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 703.210463] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061331, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.299226] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061332, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.304530] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061333, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.410888] env[61987]: DEBUG oslo_concurrency.lockutils [req-537d414c-1f2e-4c7f-b047-5ce8e1d47700 req-8ff89854-366c-469a-a8d0-ddce439ad371 service nova] Releasing lock "refresh_cache-51a36102-795e-47b7-b96a-857e54dc703e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.522172] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': task-1061326, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.526395] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Acquiring lock "8cf7f153-961a-4fab-9dd3-7b7feb2f51c1" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.527053] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Lock "8cf7f153-961a-4fab-9dd3-7b7feb2f51c1" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.710216] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061331, 'name': CreateVM_Task, 'duration_secs': 0.585231} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.710409] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 703.711197] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.711480] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.712172] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 703.712590] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1020dab5-2ab1-4969-9693-c673057d3b93 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.717606] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 703.717606] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]523a31a9-dc76-ceba-fcff-4824b29af792" [ 703.717606] env[61987]: _type = "Task" [ 703.717606] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.725999] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523a31a9-dc76-ceba-fcff-4824b29af792, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.797990] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061332, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.803024] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061333, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.967105] env[61987]: DEBUG nova.network.neutron [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Successfully updated port: 5134697b-ec80-40fc-9a81-5f002c9421bc {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 704.024207] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': task-1061326, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.028920] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Lock "8cf7f153-961a-4fab-9dd3-7b7feb2f51c1" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.502s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.029340] env[61987]: DEBUG nova.compute.manager [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 704.199885] env[61987]: DEBUG nova.compute.manager [req-2ddd5164-49fb-4ee1-8c4f-c4faa1a86ec7 req-ede0921b-a357-4351-9486-aabea8e61e73 service nova] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Received event network-vif-plugged-5134697b-ec80-40fc-9a81-5f002c9421bc {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 704.200386] env[61987]: DEBUG oslo_concurrency.lockutils [req-2ddd5164-49fb-4ee1-8c4f-c4faa1a86ec7 req-ede0921b-a357-4351-9486-aabea8e61e73 service nova] Acquiring lock "fa943137-caf1-4b41-b644-c9146f72f956-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.200386] env[61987]: DEBUG oslo_concurrency.lockutils [req-2ddd5164-49fb-4ee1-8c4f-c4faa1a86ec7 req-ede0921b-a357-4351-9486-aabea8e61e73 service nova] Lock "fa943137-caf1-4b41-b644-c9146f72f956-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.200537] env[61987]: DEBUG oslo_concurrency.lockutils [req-2ddd5164-49fb-4ee1-8c4f-c4faa1a86ec7 req-ede0921b-a357-4351-9486-aabea8e61e73 service nova] Lock "fa943137-caf1-4b41-b644-c9146f72f956-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.200680] env[61987]: DEBUG nova.compute.manager [req-2ddd5164-49fb-4ee1-8c4f-c4faa1a86ec7 req-ede0921b-a357-4351-9486-aabea8e61e73 service nova] [instance: fa943137-caf1-4b41-b644-c9146f72f956] No waiting events found dispatching network-vif-plugged-5134697b-ec80-40fc-9a81-5f002c9421bc {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 704.200848] env[61987]: WARNING nova.compute.manager [req-2ddd5164-49fb-4ee1-8c4f-c4faa1a86ec7 req-ede0921b-a357-4351-9486-aabea8e61e73 service nova] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Received unexpected event network-vif-plugged-5134697b-ec80-40fc-9a81-5f002c9421bc for instance with vm_state building and task_state spawning. [ 704.201034] env[61987]: DEBUG nova.compute.manager [req-2ddd5164-49fb-4ee1-8c4f-c4faa1a86ec7 req-ede0921b-a357-4351-9486-aabea8e61e73 service nova] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Received event network-changed-5134697b-ec80-40fc-9a81-5f002c9421bc {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 704.201207] env[61987]: DEBUG nova.compute.manager [req-2ddd5164-49fb-4ee1-8c4f-c4faa1a86ec7 req-ede0921b-a357-4351-9486-aabea8e61e73 service nova] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Refreshing instance network info cache due to event network-changed-5134697b-ec80-40fc-9a81-5f002c9421bc. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 704.201445] env[61987]: DEBUG oslo_concurrency.lockutils [req-2ddd5164-49fb-4ee1-8c4f-c4faa1a86ec7 req-ede0921b-a357-4351-9486-aabea8e61e73 service nova] Acquiring lock "refresh_cache-fa943137-caf1-4b41-b644-c9146f72f956" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.201594] env[61987]: DEBUG oslo_concurrency.lockutils [req-2ddd5164-49fb-4ee1-8c4f-c4faa1a86ec7 req-ede0921b-a357-4351-9486-aabea8e61e73 service nova] Acquired lock "refresh_cache-fa943137-caf1-4b41-b644-c9146f72f956" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.201768] env[61987]: DEBUG nova.network.neutron [req-2ddd5164-49fb-4ee1-8c4f-c4faa1a86ec7 req-ede0921b-a357-4351-9486-aabea8e61e73 service nova] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Refreshing network info cache for port 5134697b-ec80-40fc-9a81-5f002c9421bc {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 704.228264] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523a31a9-dc76-ceba-fcff-4824b29af792, 'name': SearchDatastore_Task, 'duration_secs': 0.025855} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.231012] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.231208] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 704.231399] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.231552] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.231736] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 704.232202] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0c75f740-7c5e-4b1f-9312-83a549157d01 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.246606] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 704.246819] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 704.247594] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42b452c5-a127-4fa2-b388-fa03c0e367c3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.253763] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 704.253763] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]520c722b-a3a8-387b-4d93-fb2f18dbaf21" [ 704.253763] env[61987]: _type = "Task" [ 704.253763] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.265634] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]520c722b-a3a8-387b-4d93-fb2f18dbaf21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.300847] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061332, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.306732] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061333, 'name': PowerOnVM_Task} progress is 71%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.427147] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4885b986-9a11-428b-9fcb-660da34683c0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.435531] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d97d36a-8a41-4920-b028-50f26eb68a59 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.468952] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef5b2f06-12e1-4bae-8a77-cda594e3dc0a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.472094] env[61987]: DEBUG oslo_concurrency.lockutils [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Acquiring lock "refresh_cache-fa943137-caf1-4b41-b644-c9146f72f956" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.479803] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaf263af-7acd-4db3-a164-b43de5363404 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.494478] env[61987]: DEBUG nova.compute.provider_tree [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.524230] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': task-1061326, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.21198} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.524596] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 0b4b3bdb-7393-47ab-abc1-44c47496cb3a/0b4b3bdb-7393-47ab-abc1-44c47496cb3a.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 704.524839] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 704.525230] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5590280f-ccb8-4fda-84ff-101a273fdc56 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.533212] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Waiting for the task: (returnval){ [ 704.533212] env[61987]: value = "task-1061334" [ 704.533212] env[61987]: _type = "Task" [ 704.533212] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.537549] env[61987]: DEBUG nova.compute.utils [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 704.539209] env[61987]: DEBUG nova.compute.manager [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 704.539398] env[61987]: DEBUG nova.network.neutron [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 704.546397] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': task-1061334, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.618070] env[61987]: DEBUG nova.policy [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '000274628418465a8783292646495966', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f7c49bb928a840fca8822c04190787b4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 704.756171] env[61987]: DEBUG nova.network.neutron [req-2ddd5164-49fb-4ee1-8c4f-c4faa1a86ec7 req-ede0921b-a357-4351-9486-aabea8e61e73 service nova] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 704.768327] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]520c722b-a3a8-387b-4d93-fb2f18dbaf21, 'name': SearchDatastore_Task, 'duration_secs': 0.016058} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.769302] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1243c20-6077-417b-b345-473b0a7dc800 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.776588] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 704.776588] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]521e4269-7f3f-82ab-de2e-793dfa0084b3" [ 704.776588] env[61987]: _type = "Task" [ 704.776588] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.784323] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]521e4269-7f3f-82ab-de2e-793dfa0084b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.796883] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061332, 'name': ReconfigVM_Task, 'duration_secs': 1.543087} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.801612] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Reconfigured VM instance instance-00000024 to attach disk [datastore1] 02fe490b-75c8-406b-98e0-aa200f249673/02fe490b-75c8-406b-98e0-aa200f249673.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 704.802283] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f6f91850-1f03-42c0-8784-25688887bbd6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.809305] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061333, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.810595] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 704.810595] env[61987]: value = "task-1061335" [ 704.810595] env[61987]: _type = "Task" [ 704.810595] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.818611] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061335, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.834617] env[61987]: DEBUG nova.network.neutron [req-2ddd5164-49fb-4ee1-8c4f-c4faa1a86ec7 req-ede0921b-a357-4351-9486-aabea8e61e73 service nova] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.991993] env[61987]: DEBUG nova.network.neutron [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Successfully created port: af8b91f3-b342-4408-b3da-5489a317c8b0 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 704.998113] env[61987]: DEBUG nova.scheduler.client.report [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 705.043647] env[61987]: DEBUG nova.compute.manager [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 705.046407] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': task-1061334, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.194072} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.046879] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 705.047750] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bde63ca-5959-4da4-90f8-ec9458571361 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.071052] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Reconfiguring VM instance instance-00000025 to attach disk [datastore1] 0b4b3bdb-7393-47ab-abc1-44c47496cb3a/0b4b3bdb-7393-47ab-abc1-44c47496cb3a.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 705.071393] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9b6cb89-b24b-4891-8525-df8ff17ca7aa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.090856] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Waiting for the task: (returnval){ [ 705.090856] env[61987]: value = "task-1061336" [ 705.090856] env[61987]: _type = "Task" [ 705.090856] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.100790] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': task-1061336, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.289247] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]521e4269-7f3f-82ab-de2e-793dfa0084b3, 'name': SearchDatastore_Task, 'duration_secs': 0.011187} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.289581] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.289874] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 51a36102-795e-47b7-b96a-857e54dc703e/51a36102-795e-47b7-b96a-857e54dc703e.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 705.290190] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9ed9528b-cb53-49f2-93c1-5a80d1e81b6b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.303398] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 705.303398] env[61987]: value = "task-1061337" [ 705.303398] env[61987]: _type = "Task" [ 705.303398] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.310163] env[61987]: DEBUG oslo_vmware.api [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061333, 'name': PowerOnVM_Task, 'duration_secs': 2.17174} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.311948] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 705.312127] env[61987]: INFO nova.compute.manager [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Took 13.01 seconds to spawn the instance on the hypervisor. [ 705.312338] env[61987]: DEBUG nova.compute.manager [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 705.314929] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f25eb0dd-fe09-46c1-8338-1168ccd04257 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.323634] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061337, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.329032] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061335, 'name': Rename_Task, 'duration_secs': 0.176741} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.330516] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 705.333334] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-775361b7-1485-4c4d-9af0-7fc74223ceaf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.336812] env[61987]: DEBUG oslo_concurrency.lockutils [req-2ddd5164-49fb-4ee1-8c4f-c4faa1a86ec7 req-ede0921b-a357-4351-9486-aabea8e61e73 service nova] Releasing lock "refresh_cache-fa943137-caf1-4b41-b644-c9146f72f956" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.337164] env[61987]: DEBUG oslo_concurrency.lockutils [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Acquired lock "refresh_cache-fa943137-caf1-4b41-b644-c9146f72f956" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.337321] env[61987]: DEBUG nova.network.neutron [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 705.340730] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 705.340730] env[61987]: value = "task-1061338" [ 705.340730] env[61987]: _type = "Task" [ 705.340730] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.351413] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061338, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.504659] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.504659] env[61987]: DEBUG nova.compute.manager [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 705.506556] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.269s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.507978] env[61987]: INFO nova.compute.claims [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 705.606291] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': task-1061336, 'name': ReconfigVM_Task, 'duration_secs': 0.398667} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.606660] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Reconfigured VM instance instance-00000025 to attach disk [datastore1] 0b4b3bdb-7393-47ab-abc1-44c47496cb3a/0b4b3bdb-7393-47ab-abc1-44c47496cb3a.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 705.607388] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9c09c906-bd03-4e09-b121-11ffbf2d13c5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.617267] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Waiting for the task: (returnval){ [ 705.617267] env[61987]: value = "task-1061339" [ 705.617267] env[61987]: _type = "Task" [ 705.617267] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.629299] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': task-1061339, 'name': Rename_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.814815] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061337, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.852114] env[61987]: INFO nova.compute.manager [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Took 47.00 seconds to build instance. [ 705.860351] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061338, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.899756] env[61987]: DEBUG nova.network.neutron [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 706.012194] env[61987]: DEBUG nova.compute.utils [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 706.023137] env[61987]: DEBUG nova.compute.manager [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 706.024052] env[61987]: DEBUG nova.network.neutron [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 706.055224] env[61987]: DEBUG nova.compute.manager [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 706.096288] env[61987]: DEBUG nova.policy [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8e6015b22d34635806fc28cc098f75b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ea973b343034355916bca2cce20a22f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 706.099993] env[61987]: DEBUG nova.virt.hardware [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 706.100304] env[61987]: DEBUG nova.virt.hardware [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 706.100519] env[61987]: DEBUG nova.virt.hardware [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 706.100775] env[61987]: DEBUG nova.virt.hardware [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 706.100972] env[61987]: DEBUG nova.virt.hardware [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 706.101199] env[61987]: DEBUG nova.virt.hardware [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 706.101641] env[61987]: DEBUG nova.virt.hardware [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 706.101866] env[61987]: DEBUG nova.virt.hardware [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 706.102106] env[61987]: DEBUG nova.virt.hardware [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 706.102339] env[61987]: DEBUG nova.virt.hardware [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 706.102567] env[61987]: DEBUG nova.virt.hardware [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 706.104146] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6f564e6-3216-4f1c-b688-85c598ca0596 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.117373] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed681bc2-77f2-41bd-b096-8fa492b51e7f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.131435] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': task-1061339, 'name': Rename_Task, 'duration_secs': 0.277261} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.142615] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 706.143371] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-846f695b-fa10-440f-81a1-aea79502360b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.149646] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Waiting for the task: (returnval){ [ 706.149646] env[61987]: value = "task-1061340" [ 706.149646] env[61987]: _type = "Task" [ 706.149646] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.158729] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': task-1061340, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.162602] env[61987]: DEBUG nova.network.neutron [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Updating instance_info_cache with network_info: [{"id": "5134697b-ec80-40fc-9a81-5f002c9421bc", "address": "fa:16:3e:a0:76:d2", "network": {"id": "8a9770a4-fed3-4232-ac78-8a1fc2223229", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.142", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "705320a8b6a549d0b197ee1a35404d52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bafe8721-91d4-4127-b215-d9e8e27947dc", "external-id": "nsx-vlan-transportzone-680", "segmentation_id": 680, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5134697b-ec", "ovs_interfaceid": "5134697b-ec80-40fc-9a81-5f002c9421bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.314325] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061337, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.554496} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.317553] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 51a36102-795e-47b7-b96a-857e54dc703e/51a36102-795e-47b7-b96a-857e54dc703e.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 706.317553] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 706.317831] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-112db603-0201-44bb-ae10-b4655fa3bb68 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.328025] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 706.328025] env[61987]: value = "task-1061341" [ 706.328025] env[61987]: _type = "Task" [ 706.328025] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.338919] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061341, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.356972] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d7f3d918-ec0b-4c38-b7ee-50f990c7bc4b tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "ea6c2120-a69a-42c7-bc92-3e74539c008d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.852s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.357188] env[61987]: DEBUG oslo_vmware.api [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061338, 'name': PowerOnVM_Task, 'duration_secs': 0.673857} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.358904] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 706.359106] env[61987]: INFO nova.compute.manager [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Took 11.57 seconds to spawn the instance on the hypervisor. [ 706.359398] env[61987]: DEBUG nova.compute.manager [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 706.361139] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2de1f978-c4b9-48ee-a661-3551b09cccf9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.462174] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3e11411-3e71-4f3a-b910-4b19d6fd5f1c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.471730] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-819cf356-c322-4479-bda9-9818dd0f17b5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.507608] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c39f311-6f48-44c1-a108-9c27566d229c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.516302] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8f66399-9ca7-41e2-8a53-0db22b6536a9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.521349] env[61987]: DEBUG nova.compute.manager [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 706.536711] env[61987]: DEBUG nova.compute.provider_tree [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 706.628649] env[61987]: DEBUG nova.network.neutron [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Successfully created port: fd844030-513e-4774-a178-8a6203365b9c {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 706.664375] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': task-1061340, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.664955] env[61987]: DEBUG oslo_concurrency.lockutils [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Releasing lock "refresh_cache-fa943137-caf1-4b41-b644-c9146f72f956" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.665288] env[61987]: DEBUG nova.compute.manager [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Instance network_info: |[{"id": "5134697b-ec80-40fc-9a81-5f002c9421bc", "address": "fa:16:3e:a0:76:d2", "network": {"id": "8a9770a4-fed3-4232-ac78-8a1fc2223229", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.142", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "705320a8b6a549d0b197ee1a35404d52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bafe8721-91d4-4127-b215-d9e8e27947dc", "external-id": "nsx-vlan-transportzone-680", "segmentation_id": 680, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5134697b-ec", "ovs_interfaceid": "5134697b-ec80-40fc-9a81-5f002c9421bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 706.665682] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:76:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bafe8721-91d4-4127-b215-d9e8e27947dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5134697b-ec80-40fc-9a81-5f002c9421bc', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 706.681430] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Creating folder: Project (bbd655fdd1f54787bf2056369899c924). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 706.681948] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-73af40ba-81bf-4103-928f-b81fe1b12f9d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.695333] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Created folder: Project (bbd655fdd1f54787bf2056369899c924) in parent group-v234219. [ 706.695662] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Creating folder: Instances. Parent ref: group-v234256. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 706.696014] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bf77d672-7ff3-4cd2-b45b-1b5f7b5ba57f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.706017] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Created folder: Instances in parent group-v234256. [ 706.706336] env[61987]: DEBUG oslo.service.loopingcall [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 706.706598] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 706.706837] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-353df761-1038-4e51-b7af-4c99c305e0d7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.729566] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 706.729566] env[61987]: value = "task-1061344" [ 706.729566] env[61987]: _type = "Task" [ 706.729566] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.739075] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061344, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.836834] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061341, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074966} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.837163] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 706.837978] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79de1872-4824-4beb-9824-d097f630c592 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.860730] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Reconfiguring VM instance instance-00000026 to attach disk [datastore2] 51a36102-795e-47b7-b96a-857e54dc703e/51a36102-795e-47b7-b96a-857e54dc703e.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 706.861622] env[61987]: DEBUG nova.compute.manager [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 706.864288] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b35ea68a-dd55-45cc-965d-113b9138927c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.892021] env[61987]: INFO nova.compute.manager [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Took 46.06 seconds to build instance. [ 706.892835] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 706.892835] env[61987]: value = "task-1061345" [ 706.892835] env[61987]: _type = "Task" [ 706.892835] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.905389] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061345, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.994285] env[61987]: DEBUG nova.compute.manager [req-7facacd5-f0a9-40d9-8de2-5b8b91df1471 req-840d524a-516b-4165-920e-1a52232819ac service nova] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Received event network-vif-plugged-af8b91f3-b342-4408-b3da-5489a317c8b0 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 706.994547] env[61987]: DEBUG oslo_concurrency.lockutils [req-7facacd5-f0a9-40d9-8de2-5b8b91df1471 req-840d524a-516b-4165-920e-1a52232819ac service nova] Acquiring lock "49e0d752-6f88-45c4-bd6a-d9fe4f526a2f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.994794] env[61987]: DEBUG oslo_concurrency.lockutils [req-7facacd5-f0a9-40d9-8de2-5b8b91df1471 req-840d524a-516b-4165-920e-1a52232819ac service nova] Lock "49e0d752-6f88-45c4-bd6a-d9fe4f526a2f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.995014] env[61987]: DEBUG oslo_concurrency.lockutils [req-7facacd5-f0a9-40d9-8de2-5b8b91df1471 req-840d524a-516b-4165-920e-1a52232819ac service nova] Lock "49e0d752-6f88-45c4-bd6a-d9fe4f526a2f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.995515] env[61987]: DEBUG nova.compute.manager [req-7facacd5-f0a9-40d9-8de2-5b8b91df1471 req-840d524a-516b-4165-920e-1a52232819ac service nova] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] No waiting events found dispatching network-vif-plugged-af8b91f3-b342-4408-b3da-5489a317c8b0 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 706.995702] env[61987]: WARNING nova.compute.manager [req-7facacd5-f0a9-40d9-8de2-5b8b91df1471 req-840d524a-516b-4165-920e-1a52232819ac service nova] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Received unexpected event network-vif-plugged-af8b91f3-b342-4408-b3da-5489a317c8b0 for instance with vm_state building and task_state spawning. [ 707.035549] env[61987]: DEBUG nova.network.neutron [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Successfully updated port: af8b91f3-b342-4408-b3da-5489a317c8b0 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 707.074484] env[61987]: ERROR nova.scheduler.client.report [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [req-6a4c91a9-0985-4fda-aa47-77871844d98a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8db744e5-1e14-4e98-9818-695ea4b7dfd0. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-6a4c91a9-0985-4fda-aa47-77871844d98a"}]} [ 707.099586] env[61987]: DEBUG nova.scheduler.client.report [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Refreshing inventories for resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 707.126301] env[61987]: DEBUG nova.scheduler.client.report [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Updating ProviderTree inventory for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 707.126301] env[61987]: DEBUG nova.compute.provider_tree [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 707.145787] env[61987]: DEBUG nova.scheduler.client.report [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Refreshing aggregate associations for resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0, aggregates: None {{(pid=61987) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 707.162612] env[61987]: DEBUG oslo_vmware.api [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': task-1061340, 'name': PowerOnVM_Task, 'duration_secs': 0.584018} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.162973] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 707.163232] env[61987]: INFO nova.compute.manager [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Took 9.89 seconds to spawn the instance on the hypervisor. [ 707.163500] env[61987]: DEBUG nova.compute.manager [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 707.164487] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fad9f83-a58f-4441-92c8-826c982f351b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.177429] env[61987]: DEBUG nova.scheduler.client.report [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Refreshing trait associations for resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61987) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 707.240967] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061344, 'name': CreateVM_Task, 'duration_secs': 0.466411} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.241180] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 707.241906] env[61987]: DEBUG oslo_concurrency.lockutils [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.242098] env[61987]: DEBUG oslo_concurrency.lockutils [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.243181] env[61987]: DEBUG oslo_concurrency.lockutils [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 707.243181] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91d31534-0040-4a3b-a9d2-30408c251026 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.247889] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Waiting for the task: (returnval){ [ 707.247889] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52eb41c7-35b9-b870-d263-14419610b73d" [ 707.247889] env[61987]: _type = "Task" [ 707.247889] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.261727] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52eb41c7-35b9-b870-d263-14419610b73d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.396664] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2168a6ff-a1fc-432c-84ba-69667e65bc87 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "02fe490b-75c8-406b-98e0-aa200f249673" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.386s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.408092] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061345, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.411812] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.532526] env[61987]: DEBUG nova.compute.manager [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 707.539708] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Acquiring lock "refresh_cache-49e0d752-6f88-45c4-bd6a-d9fe4f526a2f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.539888] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Acquired lock "refresh_cache-49e0d752-6f88-45c4-bd6a-d9fe4f526a2f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.540066] env[61987]: DEBUG nova.network.neutron [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 707.582837] env[61987]: DEBUG nova.virt.hardware [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 707.583204] env[61987]: DEBUG nova.virt.hardware [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 707.583306] env[61987]: DEBUG nova.virt.hardware [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 707.583529] env[61987]: DEBUG nova.virt.hardware [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 707.583635] env[61987]: DEBUG nova.virt.hardware [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 707.583773] env[61987]: DEBUG nova.virt.hardware [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 707.583986] env[61987]: DEBUG nova.virt.hardware [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 707.584170] env[61987]: DEBUG nova.virt.hardware [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 707.584344] env[61987]: DEBUG nova.virt.hardware [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 707.584516] env[61987]: DEBUG nova.virt.hardware [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 707.584720] env[61987]: DEBUG nova.virt.hardware [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 707.585617] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2867458-23c8-44ec-8b06-b993cb88814d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.596357] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e1289c-1e29-4123-9e92-d5b6d0b12f7c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.644399] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77badbf8-051e-4637-99aa-99022635b66d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.652363] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d271222c-92b1-448c-b797-0efc1c64c708 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.691699] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f972d8c-5946-49e9-9d34-4ad58430211c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.697765] env[61987]: INFO nova.compute.manager [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Took 44.86 seconds to build instance. [ 707.702564] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74aa4169-ae7b-4b09-90d0-e61a4ea792b1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.719878] env[61987]: DEBUG nova.compute.provider_tree [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 707.759516] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52eb41c7-35b9-b870-d263-14419610b73d, 'name': SearchDatastore_Task, 'duration_secs': 0.017027} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.759745] env[61987]: DEBUG oslo_concurrency.lockutils [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.759976] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 707.760265] env[61987]: DEBUG oslo_concurrency.lockutils [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.760424] env[61987]: DEBUG oslo_concurrency.lockutils [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.760628] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 707.760921] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4c79be96-28e3-484f-9f31-3f62756a0468 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.770497] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 707.770894] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 707.772017] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d64bbb16-25fb-4e2c-bdcb-2cfaea9a75b6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.778937] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Waiting for the task: (returnval){ [ 707.778937] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52a5c5ac-3dd6-99b9-470f-75ecd472d3ac" [ 707.778937] env[61987]: _type = "Task" [ 707.778937] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.791226] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52a5c5ac-3dd6-99b9-470f-75ecd472d3ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.902338] env[61987]: DEBUG nova.compute.manager [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 707.911769] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061345, 'name': ReconfigVM_Task, 'duration_secs': 0.708232} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.912157] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Reconfigured VM instance instance-00000026 to attach disk [datastore2] 51a36102-795e-47b7-b96a-857e54dc703e/51a36102-795e-47b7-b96a-857e54dc703e.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 707.912843] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8bf6d9b1-305f-4e0d-9c64-be3bdbe45db3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.919933] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 707.919933] env[61987]: value = "task-1061346" [ 707.919933] env[61987]: _type = "Task" [ 707.919933] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.929581] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061346, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.080493] env[61987]: DEBUG nova.network.neutron [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 708.202253] env[61987]: DEBUG oslo_concurrency.lockutils [None req-25a227ff-ccae-43d5-9c72-ca903ba35b71 tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Lock "0b4b3bdb-7393-47ab-abc1-44c47496cb3a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.284s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.278403] env[61987]: DEBUG nova.scheduler.client.report [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Updated inventory for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with generation 63 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 708.278672] env[61987]: DEBUG nova.compute.provider_tree [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Updating resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 generation from 63 to 64 during operation: update_inventory {{(pid=61987) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 708.279036] env[61987]: DEBUG nova.compute.provider_tree [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 708.291967] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52a5c5ac-3dd6-99b9-470f-75ecd472d3ac, 'name': SearchDatastore_Task, 'duration_secs': 0.014075} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.292866] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39de9e4d-7c5c-4139-b45f-8431f8f879f3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.298932] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Waiting for the task: (returnval){ [ 708.298932] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52fed4b7-1756-d626-2573-2f68d1800bd9" [ 708.298932] env[61987]: _type = "Task" [ 708.298932] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.306956] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52fed4b7-1756-d626-2573-2f68d1800bd9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.416531] env[61987]: DEBUG nova.network.neutron [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Updating instance_info_cache with network_info: [{"id": "af8b91f3-b342-4408-b3da-5489a317c8b0", "address": "fa:16:3e:ca:db:34", "network": {"id": "1694a2d2-e9f7-4e79-b1d2-a5d9bd32b72d", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-980788300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7c49bb928a840fca8822c04190787b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e39ca24f-7890-4cdf-8dab-ecab218bb063", "external-id": "nsx-vlan-transportzone-890", "segmentation_id": 890, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf8b91f3-b3", "ovs_interfaceid": "af8b91f3-b342-4408-b3da-5489a317c8b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.433483] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061346, 'name': Rename_Task, 'duration_secs': 0.313624} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.433950] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 708.434349] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-693dfd3f-fde8-45fd-aed5-72df6834dc62 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.445254] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 708.445254] env[61987]: value = "task-1061347" [ 708.445254] env[61987]: _type = "Task" [ 708.445254] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.454893] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061347, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.457155] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.570082] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-260745b7-a40f-48ec-81cd-cb2261a3727e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.577116] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-403357ab-3ae8-4209-b36d-22f389366461 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Suspending the VM {{(pid=61987) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 708.577431] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-a3e2e8d4-bd77-409b-9114-2910b9aa4a7f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.584151] env[61987]: DEBUG oslo_vmware.api [None req-403357ab-3ae8-4209-b36d-22f389366461 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 708.584151] env[61987]: value = "task-1061348" [ 708.584151] env[61987]: _type = "Task" [ 708.584151] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.593038] env[61987]: DEBUG oslo_vmware.api [None req-403357ab-3ae8-4209-b36d-22f389366461 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061348, 'name': SuspendVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.705154] env[61987]: DEBUG nova.compute.manager [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 708.720429] env[61987]: DEBUG nova.network.neutron [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Successfully updated port: fd844030-513e-4774-a178-8a6203365b9c {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 708.789865] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.282s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.789932] env[61987]: DEBUG nova.compute.manager [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 708.793382] env[61987]: DEBUG oslo_concurrency.lockutils [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.189s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.798719] env[61987]: INFO nova.compute.claims [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 708.820638] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52fed4b7-1756-d626-2573-2f68d1800bd9, 'name': SearchDatastore_Task, 'duration_secs': 0.030152} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.820638] env[61987]: DEBUG oslo_concurrency.lockutils [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.820638] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] fa943137-caf1-4b41-b644-c9146f72f956/fa943137-caf1-4b41-b644-c9146f72f956.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 708.820638] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d961ec02-cb2b-46bc-a791-5ce20d978858 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.829055] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Waiting for the task: (returnval){ [ 708.829055] env[61987]: value = "task-1061349" [ 708.829055] env[61987]: _type = "Task" [ 708.829055] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.840492] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': task-1061349, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.920978] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Releasing lock "refresh_cache-49e0d752-6f88-45c4-bd6a-d9fe4f526a2f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.920978] env[61987]: DEBUG nova.compute.manager [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Instance network_info: |[{"id": "af8b91f3-b342-4408-b3da-5489a317c8b0", "address": "fa:16:3e:ca:db:34", "network": {"id": "1694a2d2-e9f7-4e79-b1d2-a5d9bd32b72d", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-980788300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7c49bb928a840fca8822c04190787b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e39ca24f-7890-4cdf-8dab-ecab218bb063", "external-id": "nsx-vlan-transportzone-890", "segmentation_id": 890, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf8b91f3-b3", "ovs_interfaceid": "af8b91f3-b342-4408-b3da-5489a317c8b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 708.921313] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ca:db:34', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e39ca24f-7890-4cdf-8dab-ecab218bb063', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'af8b91f3-b342-4408-b3da-5489a317c8b0', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 708.929807] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Creating folder: Project (f7c49bb928a840fca8822c04190787b4). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 708.930472] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-04d7f577-cf02-4c97-8f80-0c8c5191e5ac {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.942779] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Created folder: Project (f7c49bb928a840fca8822c04190787b4) in parent group-v234219. [ 708.943037] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Creating folder: Instances. Parent ref: group-v234259. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 708.943320] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b3d43980-a29f-4882-b244-82449fa39497 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.953724] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061347, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.955728] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Created folder: Instances in parent group-v234259. [ 708.955994] env[61987]: DEBUG oslo.service.loopingcall [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 708.956213] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 708.956598] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ac4808df-710b-42f8-ac74-e5073bc8d947 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.975979] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 708.975979] env[61987]: value = "task-1061352" [ 708.975979] env[61987]: _type = "Task" [ 708.975979] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.985045] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061352, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.097304] env[61987]: DEBUG oslo_vmware.api [None req-403357ab-3ae8-4209-b36d-22f389366461 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061348, 'name': SuspendVM_Task} progress is 50%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.223579] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Acquiring lock "refresh_cache-8ff4ebb9-57ea-49d2-8d3c-5876da61e77b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.224303] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Acquired lock "refresh_cache-8ff4ebb9-57ea-49d2-8d3c-5876da61e77b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.224303] env[61987]: DEBUG nova.network.neutron [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 709.227124] env[61987]: DEBUG nova.compute.manager [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Received event network-changed-af8b91f3-b342-4408-b3da-5489a317c8b0 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 709.227365] env[61987]: DEBUG nova.compute.manager [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Refreshing instance network info cache due to event network-changed-af8b91f3-b342-4408-b3da-5489a317c8b0. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 709.227530] env[61987]: DEBUG oslo_concurrency.lockutils [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] Acquiring lock "refresh_cache-49e0d752-6f88-45c4-bd6a-d9fe4f526a2f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.227675] env[61987]: DEBUG oslo_concurrency.lockutils [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] Acquired lock "refresh_cache-49e0d752-6f88-45c4-bd6a-d9fe4f526a2f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.227842] env[61987]: DEBUG nova.network.neutron [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Refreshing network info cache for port af8b91f3-b342-4408-b3da-5489a317c8b0 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 709.250818] env[61987]: DEBUG oslo_concurrency.lockutils [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.300878] env[61987]: DEBUG nova.compute.utils [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 709.305631] env[61987]: DEBUG nova.compute.manager [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 709.305935] env[61987]: DEBUG nova.network.neutron [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 709.338731] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': task-1061349, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.366754] env[61987]: DEBUG nova.policy [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd12c31d96c9846b496606690b4f55ff7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8c3057c923f84bc1a2b00e9526736e1d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 709.456250] env[61987]: DEBUG oslo_vmware.api [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061347, 'name': PowerOnVM_Task, 'duration_secs': 0.632152} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.456535] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 709.456700] env[61987]: INFO nova.compute.manager [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Took 9.55 seconds to spawn the instance on the hypervisor. [ 709.456882] env[61987]: DEBUG nova.compute.manager [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 709.457764] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aca38dd1-390e-4472-9d59-b13141b4e164 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.487027] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061352, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.595166] env[61987]: DEBUG oslo_vmware.api [None req-403357ab-3ae8-4209-b36d-22f389366461 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061348, 'name': SuspendVM_Task, 'duration_secs': 0.797057} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.595474] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-403357ab-3ae8-4209-b36d-22f389366461 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Suspended the VM {{(pid=61987) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 709.595676] env[61987]: DEBUG nova.compute.manager [None req-403357ab-3ae8-4209-b36d-22f389366461 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 709.596540] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-762fd572-8c9b-4ea0-929c-7837329fb4aa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.809635] env[61987]: DEBUG nova.compute.manager [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 709.820704] env[61987]: DEBUG nova.network.neutron [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 709.837721] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': task-1061349, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.565971} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.838222] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] fa943137-caf1-4b41-b644-c9146f72f956/fa943137-caf1-4b41-b644-c9146f72f956.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 709.838512] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 709.838871] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7cdb3cc4-b131-4e12-a86f-1eb5ac45241c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.850175] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Waiting for the task: (returnval){ [ 709.850175] env[61987]: value = "task-1061353" [ 709.850175] env[61987]: _type = "Task" [ 709.850175] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.869301] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': task-1061353, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.911400] env[61987]: DEBUG nova.network.neutron [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Successfully created port: f0fc4290-34ae-42ea-9704-e52bdcfdee85 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 709.977605] env[61987]: INFO nova.compute.manager [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Took 45.02 seconds to build instance. [ 709.987374] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061352, 'name': CreateVM_Task, 'duration_secs': 0.534436} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.987544] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 709.988272] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.988456] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.988796] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 709.989062] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fea2f603-d7ab-4cd4-ab30-03c4745b4bfe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.996578] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Waiting for the task: (returnval){ [ 709.996578] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52d52446-f33c-4dd0-0507-cd1914e4b442" [ 709.996578] env[61987]: _type = "Task" [ 709.996578] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.003683] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52d52446-f33c-4dd0-0507-cd1914e4b442, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.218052] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d71d63-bf05-4766-a3b8-b71242946bbd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.228260] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ca45d96-7719-4af2-94d6-0069b8edeaa5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.261820] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b390afee-b119-4edf-ac98-595d8bca6acd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.271207] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfebc022-f8af-46a2-8653-61675eda5e62 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.285706] env[61987]: DEBUG nova.compute.provider_tree [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 710.295226] env[61987]: DEBUG nova.network.neutron [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Updating instance_info_cache with network_info: [{"id": "fd844030-513e-4774-a178-8a6203365b9c", "address": "fa:16:3e:a8:bd:b9", "network": {"id": "662d2418-580e-4675-974e-8fca3f56aaa1", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1448080498-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ea973b343034355916bca2cce20a22f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd844030-51", "ovs_interfaceid": "fd844030-513e-4774-a178-8a6203365b9c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.364846] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': task-1061353, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064583} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.365340] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 710.366891] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197742e0-0427-474b-be53-3305aec1762c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.372836] env[61987]: DEBUG nova.network.neutron [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Updated VIF entry in instance network info cache for port af8b91f3-b342-4408-b3da-5489a317c8b0. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 710.373547] env[61987]: DEBUG nova.network.neutron [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Updating instance_info_cache with network_info: [{"id": "af8b91f3-b342-4408-b3da-5489a317c8b0", "address": "fa:16:3e:ca:db:34", "network": {"id": "1694a2d2-e9f7-4e79-b1d2-a5d9bd32b72d", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-980788300-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7c49bb928a840fca8822c04190787b4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e39ca24f-7890-4cdf-8dab-ecab218bb063", "external-id": "nsx-vlan-transportzone-890", "segmentation_id": 890, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf8b91f3-b3", "ovs_interfaceid": "af8b91f3-b342-4408-b3da-5489a317c8b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.399311] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] fa943137-caf1-4b41-b644-c9146f72f956/fa943137-caf1-4b41-b644-c9146f72f956.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 710.400180] env[61987]: DEBUG oslo_concurrency.lockutils [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] Releasing lock "refresh_cache-49e0d752-6f88-45c4-bd6a-d9fe4f526a2f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.400563] env[61987]: DEBUG nova.compute.manager [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Received event network-vif-plugged-fd844030-513e-4774-a178-8a6203365b9c {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 710.400806] env[61987]: DEBUG oslo_concurrency.lockutils [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] Acquiring lock "8ff4ebb9-57ea-49d2-8d3c-5876da61e77b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.401128] env[61987]: DEBUG oslo_concurrency.lockutils [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] Lock "8ff4ebb9-57ea-49d2-8d3c-5876da61e77b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.401334] env[61987]: DEBUG oslo_concurrency.lockutils [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] Lock "8ff4ebb9-57ea-49d2-8d3c-5876da61e77b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.401549] env[61987]: DEBUG nova.compute.manager [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] No waiting events found dispatching network-vif-plugged-fd844030-513e-4774-a178-8a6203365b9c {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 710.401799] env[61987]: WARNING nova.compute.manager [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Received unexpected event network-vif-plugged-fd844030-513e-4774-a178-8a6203365b9c for instance with vm_state building and task_state spawning. [ 710.401962] env[61987]: DEBUG nova.compute.manager [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Received event network-changed-fd844030-513e-4774-a178-8a6203365b9c {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 710.403162] env[61987]: DEBUG nova.compute.manager [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Refreshing instance network info cache due to event network-changed-fd844030-513e-4774-a178-8a6203365b9c. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 710.403162] env[61987]: DEBUG oslo_concurrency.lockutils [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] Acquiring lock "refresh_cache-8ff4ebb9-57ea-49d2-8d3c-5876da61e77b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.403162] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f9a502e-6afc-4b37-a5b9-ccb6ccdd3de5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.425199] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Waiting for the task: (returnval){ [ 710.425199] env[61987]: value = "task-1061354" [ 710.425199] env[61987]: _type = "Task" [ 710.425199] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.433798] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': task-1061354, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.483099] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9940ebbe-f8c3-469e-af59-be3ec522c1e2 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "51a36102-795e-47b7-b96a-857e54dc703e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.673s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.512736] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52d52446-f33c-4dd0-0507-cd1914e4b442, 'name': SearchDatastore_Task, 'duration_secs': 0.021153} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.513090] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.513344] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 710.513668] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.513834] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.514109] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 710.514412] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ce6f1da8-e4ba-473e-9fd0-d696f61c590a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.524131] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 710.524324] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 710.525137] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00c707bc-e3a6-490a-9f3a-224f0af3b5da {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.530802] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Waiting for the task: (returnval){ [ 710.530802] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52eeb884-af24-2ed6-d50b-a4fbd4c2d116" [ 710.530802] env[61987]: _type = "Task" [ 710.530802] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.538865] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52eeb884-af24-2ed6-d50b-a4fbd4c2d116, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.791280] env[61987]: DEBUG nova.scheduler.client.report [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 710.798116] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Releasing lock "refresh_cache-8ff4ebb9-57ea-49d2-8d3c-5876da61e77b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.798423] env[61987]: DEBUG nova.compute.manager [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Instance network_info: |[{"id": "fd844030-513e-4774-a178-8a6203365b9c", "address": "fa:16:3e:a8:bd:b9", "network": {"id": "662d2418-580e-4675-974e-8fca3f56aaa1", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1448080498-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ea973b343034355916bca2cce20a22f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd844030-51", "ovs_interfaceid": "fd844030-513e-4774-a178-8a6203365b9c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 710.798722] env[61987]: DEBUG oslo_concurrency.lockutils [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] Acquired lock "refresh_cache-8ff4ebb9-57ea-49d2-8d3c-5876da61e77b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.798916] env[61987]: DEBUG nova.network.neutron [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Refreshing network info cache for port fd844030-513e-4774-a178-8a6203365b9c {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 710.800020] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a8:bd:b9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '49b5df12-d801-4140-8816-2fd401608c7d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fd844030-513e-4774-a178-8a6203365b9c', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 710.807526] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Creating folder: Project (7ea973b343034355916bca2cce20a22f). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 710.808563] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8dfb7329-4693-4525-9070-664091d335fa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.818927] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Created folder: Project (7ea973b343034355916bca2cce20a22f) in parent group-v234219. [ 710.819853] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Creating folder: Instances. Parent ref: group-v234262. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 710.819853] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b33fb32e-158f-4462-af31-200f9c63a2b2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.825568] env[61987]: DEBUG nova.compute.manager [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 710.828772] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Created folder: Instances in parent group-v234262. [ 710.829016] env[61987]: DEBUG oslo.service.loopingcall [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 710.829218] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 710.829421] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-289ba7a8-b599-4103-8c3f-1d6c11cd176e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.849622] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 710.849622] env[61987]: value = "task-1061357" [ 710.849622] env[61987]: _type = "Task" [ 710.849622] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.851742] env[61987]: DEBUG nova.virt.hardware [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 710.851977] env[61987]: DEBUG nova.virt.hardware [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 710.854810] env[61987]: DEBUG nova.virt.hardware [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 710.854810] env[61987]: DEBUG nova.virt.hardware [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 710.854810] env[61987]: DEBUG nova.virt.hardware [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 710.854810] env[61987]: DEBUG nova.virt.hardware [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 710.854810] env[61987]: DEBUG nova.virt.hardware [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 710.855013] env[61987]: DEBUG nova.virt.hardware [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 710.855013] env[61987]: DEBUG nova.virt.hardware [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 710.855013] env[61987]: DEBUG nova.virt.hardware [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 710.855013] env[61987]: DEBUG nova.virt.hardware [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 710.855013] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-775f0c5c-dfea-467c-9893-4b4d299fdb40 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.867312] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d528de-19da-47a5-b798-da48e017a260 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.871234] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061357, 'name': CreateVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.938024] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': task-1061354, 'name': ReconfigVM_Task, 'duration_secs': 0.313011} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.938024] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Reconfigured VM instance instance-00000027 to attach disk [datastore2] fa943137-caf1-4b41-b644-c9146f72f956/fa943137-caf1-4b41-b644-c9146f72f956.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 710.938024] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2c6a597f-8bdd-4a20-8e5d-6eefa00e57bc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.946021] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Waiting for the task: (returnval){ [ 710.946021] env[61987]: value = "task-1061358" [ 710.946021] env[61987]: _type = "Task" [ 710.946021] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.954864] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': task-1061358, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.987349] env[61987]: DEBUG nova.compute.manager [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 711.042281] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52eeb884-af24-2ed6-d50b-a4fbd4c2d116, 'name': SearchDatastore_Task, 'duration_secs': 0.017305} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.043125] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9886805f-fd37-4486-9896-afc66250cf34 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.049054] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Waiting for the task: (returnval){ [ 711.049054] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]521a2233-1659-5c84-3882-24b37db9f651" [ 711.049054] env[61987]: _type = "Task" [ 711.049054] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.057146] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]521a2233-1659-5c84-3882-24b37db9f651, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.296816] env[61987]: DEBUG oslo_concurrency.lockutils [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.503s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.298599] env[61987]: DEBUG nova.compute.manager [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 711.301595] env[61987]: DEBUG oslo_concurrency.lockutils [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.613s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.302053] env[61987]: DEBUG nova.objects.instance [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Lazy-loading 'resources' on Instance uuid 89a97f81-bcd9-49e5-bbd3-9140f6638a20 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 711.364616] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061357, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.459182] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': task-1061358, 'name': Rename_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.528074] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.559961] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]521a2233-1659-5c84-3882-24b37db9f651, 'name': SearchDatastore_Task, 'duration_secs': 0.013213} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.560599] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.560895] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f/49e0d752-6f88-45c4-bd6a-d9fe4f526a2f.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 711.561277] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c5f94216-bf97-40eb-a925-80f09234682a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.567509] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Waiting for the task: (returnval){ [ 711.567509] env[61987]: value = "task-1061359" [ 711.567509] env[61987]: _type = "Task" [ 711.567509] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.577687] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': task-1061359, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.763435] env[61987]: DEBUG nova.network.neutron [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Updated VIF entry in instance network info cache for port fd844030-513e-4774-a178-8a6203365b9c. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 711.764158] env[61987]: DEBUG nova.network.neutron [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Updating instance_info_cache with network_info: [{"id": "fd844030-513e-4774-a178-8a6203365b9c", "address": "fa:16:3e:a8:bd:b9", "network": {"id": "662d2418-580e-4675-974e-8fca3f56aaa1", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1448080498-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ea973b343034355916bca2cce20a22f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd844030-51", "ovs_interfaceid": "fd844030-513e-4774-a178-8a6203365b9c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.808736] env[61987]: DEBUG nova.compute.utils [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 711.812610] env[61987]: DEBUG nova.compute.manager [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 711.813101] env[61987]: DEBUG nova.network.neutron [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 711.867196] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061357, 'name': CreateVM_Task, 'duration_secs': 0.534762} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.867196] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 711.868687] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.868870] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.869228] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 711.869760] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb73aab0-0509-472e-93f1-65c27edfbea6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.877435] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Waiting for the task: (returnval){ [ 711.877435] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]527d3d6c-b807-88de-b06a-7b2b1b5e8330" [ 711.877435] env[61987]: _type = "Task" [ 711.877435] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.891874] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]527d3d6c-b807-88de-b06a-7b2b1b5e8330, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.895947] env[61987]: DEBUG nova.compute.manager [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 711.895947] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b82a3f-d149-4653-b6a4-f32996567018 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.948616] env[61987]: DEBUG nova.policy [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6b9b9da9b9248d3bffdb97582f0b1fe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd4e086581d9e41b197f84dac49aab72e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 711.960509] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': task-1061358, 'name': Rename_Task, 'duration_secs': 0.900836} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.963258] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 711.964329] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-42c82434-03fc-4c78-85fa-5e6514a92364 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.972544] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Waiting for the task: (returnval){ [ 711.972544] env[61987]: value = "task-1061360" [ 711.972544] env[61987]: _type = "Task" [ 711.972544] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.988273] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': task-1061360, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.081669] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': task-1061359, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.167746] env[61987]: DEBUG nova.network.neutron [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Successfully updated port: f0fc4290-34ae-42ea-9704-e52bdcfdee85 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 712.262896] env[61987]: DEBUG nova.compute.manager [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 712.266283] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98031fcf-9f7e-469f-a9a5-99ab23d206b6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.270858] env[61987]: DEBUG oslo_concurrency.lockutils [req-74ee09a7-06d5-4f2d-b0e1-e197e46051ad req-5d7f745e-c801-4638-9d75-ca929f0209db service nova] Releasing lock "refresh_cache-8ff4ebb9-57ea-49d2-8d3c-5876da61e77b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.283540] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8689dd2-15fb-4f92-b7ae-dfcca29887ce {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.291572] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a136e544-39ab-4217-af78-000dcb540efe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.327069] env[61987]: DEBUG nova.compute.manager [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 712.334035] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f593eb80-c75a-4bf5-bf51-bdca65d2fa3d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.337381] env[61987]: DEBUG nova.compute.manager [req-49defee9-a125-4f6a-8485-9d5dfb2a797f req-e30ea94f-4553-48b1-a74e-f86da2d7943a service nova] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Received event network-vif-plugged-f0fc4290-34ae-42ea-9704-e52bdcfdee85 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 712.337571] env[61987]: DEBUG oslo_concurrency.lockutils [req-49defee9-a125-4f6a-8485-9d5dfb2a797f req-e30ea94f-4553-48b1-a74e-f86da2d7943a service nova] Acquiring lock "e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.337774] env[61987]: DEBUG oslo_concurrency.lockutils [req-49defee9-a125-4f6a-8485-9d5dfb2a797f req-e30ea94f-4553-48b1-a74e-f86da2d7943a service nova] Lock "e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.337983] env[61987]: DEBUG oslo_concurrency.lockutils [req-49defee9-a125-4f6a-8485-9d5dfb2a797f req-e30ea94f-4553-48b1-a74e-f86da2d7943a service nova] Lock "e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.340023] env[61987]: DEBUG nova.compute.manager [req-49defee9-a125-4f6a-8485-9d5dfb2a797f req-e30ea94f-4553-48b1-a74e-f86da2d7943a service nova] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] No waiting events found dispatching network-vif-plugged-f0fc4290-34ae-42ea-9704-e52bdcfdee85 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 712.340023] env[61987]: WARNING nova.compute.manager [req-49defee9-a125-4f6a-8485-9d5dfb2a797f req-e30ea94f-4553-48b1-a74e-f86da2d7943a service nova] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Received unexpected event network-vif-plugged-f0fc4290-34ae-42ea-9704-e52bdcfdee85 for instance with vm_state building and task_state spawning. [ 712.345451] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d186ed-4914-4672-90f4-6de3d9ff2d93 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.360958] env[61987]: DEBUG nova.compute.provider_tree [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.386149] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]527d3d6c-b807-88de-b06a-7b2b1b5e8330, 'name': SearchDatastore_Task, 'duration_secs': 0.057786} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.386471] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.386854] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 712.386984] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.387150] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.387345] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 712.387813] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5b4d01bc-31b8-4cb6-b1c0-17b79185f2fc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.399227] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 712.399227] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 712.399227] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fae68a87-88c7-4cbb-be25-1bdfe404df8b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.407604] env[61987]: INFO nova.compute.manager [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] instance snapshotting [ 712.407604] env[61987]: WARNING nova.compute.manager [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 712.407604] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Waiting for the task: (returnval){ [ 712.407604] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52f65398-4d42-6d11-431a-6dbccbf66fa1" [ 712.407604] env[61987]: _type = "Task" [ 712.407604] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.408986] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64bd8642-ea14-46e7-9c93-e7f47a6697ac {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.437836] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47d36723-fc1c-44ca-9df1-720524dd34cb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.440958] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52f65398-4d42-6d11-431a-6dbccbf66fa1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.457026] env[61987]: DEBUG nova.network.neutron [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Successfully created port: fc1b4b12-b396-40ac-9eab-a1309f2259e4 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 712.490092] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': task-1061360, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.584871] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': task-1061359, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.576165} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.584871] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f/49e0d752-6f88-45c4-bd6a-d9fe4f526a2f.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 712.584871] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 712.584871] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6fc6dfc8-7703-4cc0-ba90-c50d8f921e1a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.591869] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Waiting for the task: (returnval){ [ 712.591869] env[61987]: value = "task-1061361" [ 712.591869] env[61987]: _type = "Task" [ 712.591869] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.600434] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': task-1061361, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.620969] env[61987]: DEBUG oslo_concurrency.lockutils [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "93a27dea-f28a-4cbf-ac37-ff41763ae18d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.621259] env[61987]: DEBUG oslo_concurrency.lockutils [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "93a27dea-f28a-4cbf-ac37-ff41763ae18d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.670063] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Acquiring lock "refresh_cache-e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.670233] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Acquired lock "refresh_cache-e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.670375] env[61987]: DEBUG nova.network.neutron [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 712.781319] env[61987]: INFO nova.compute.manager [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] instance snapshotting [ 712.788329] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c213ee95-43b6-4401-847a-99a0d0049287 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.809549] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2da905d5-ea65-46b3-8659-01ff22a60600 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.867898] env[61987]: DEBUG nova.scheduler.client.report [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 712.918429] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52f65398-4d42-6d11-431a-6dbccbf66fa1, 'name': SearchDatastore_Task, 'duration_secs': 0.029723} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.919250] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-943747c8-868f-486e-a0c9-1e08d307cccd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.926011] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Waiting for the task: (returnval){ [ 712.926011] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52cbff4d-1d44-098f-5a33-1fd070cfe9d3" [ 712.926011] env[61987]: _type = "Task" [ 712.926011] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.937437] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52cbff4d-1d44-098f-5a33-1fd070cfe9d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.950096] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Creating Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 712.950503] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-7f3cb4b6-c06f-42de-bb17-60fdd0984401 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.958390] env[61987]: DEBUG oslo_vmware.api [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 712.958390] env[61987]: value = "task-1061362" [ 712.958390] env[61987]: _type = "Task" [ 712.958390] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.969061] env[61987]: DEBUG oslo_vmware.api [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061362, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.983275] env[61987]: DEBUG oslo_vmware.api [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': task-1061360, 'name': PowerOnVM_Task, 'duration_secs': 0.536873} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.983607] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 712.983878] env[61987]: INFO nova.compute.manager [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Took 10.50 seconds to spawn the instance on the hypervisor. [ 712.984102] env[61987]: DEBUG nova.compute.manager [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 712.984978] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a9a583-cc3c-4f0e-8992-330cb0a01d4d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.101891] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': task-1061361, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.228405} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.102186] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 713.102945] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d00553-8278-43d0-8102-da02c0405f80 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.125497] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f/49e0d752-6f88-45c4-bd6a-d9fe4f526a2f.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 713.125764] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-26a7cafe-1179-49cc-97a7-f513dd62521f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.145050] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Waiting for the task: (returnval){ [ 713.145050] env[61987]: value = "task-1061363" [ 713.145050] env[61987]: _type = "Task" [ 713.145050] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.154411] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': task-1061363, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.207353] env[61987]: DEBUG nova.network.neutron [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 713.323890] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Creating Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 713.324256] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-dcec2274-766f-4159-8cb7-042262f6801f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.333338] env[61987]: DEBUG oslo_vmware.api [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 713.333338] env[61987]: value = "task-1061364" [ 713.333338] env[61987]: _type = "Task" [ 713.333338] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.344049] env[61987]: DEBUG nova.compute.manager [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 713.346318] env[61987]: DEBUG oslo_vmware.api [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061364, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.372194] env[61987]: DEBUG oslo_concurrency.lockutils [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.071s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.375838] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.210s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.376669] env[61987]: INFO nova.compute.claims [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 713.405080] env[61987]: DEBUG nova.virt.hardware [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:08:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='3ca61d0c-b87e-4949-982d-be5132df125d',id=36,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-2078042112',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 713.405080] env[61987]: DEBUG nova.virt.hardware [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 713.405080] env[61987]: DEBUG nova.virt.hardware [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 713.405516] env[61987]: DEBUG nova.virt.hardware [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 713.405516] env[61987]: DEBUG nova.virt.hardware [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 713.405516] env[61987]: DEBUG nova.virt.hardware [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 713.405516] env[61987]: DEBUG nova.virt.hardware [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 713.405516] env[61987]: DEBUG nova.virt.hardware [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 713.405664] env[61987]: DEBUG nova.virt.hardware [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 713.405664] env[61987]: DEBUG nova.virt.hardware [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 713.405664] env[61987]: DEBUG nova.virt.hardware [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 713.408218] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3bae50-2857-487a-995b-747301740e97 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.417205] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea6f97b5-5729-4408-8dcc-2f71e2b74ff7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.436038] env[61987]: DEBUG nova.network.neutron [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Updating instance_info_cache with network_info: [{"id": "f0fc4290-34ae-42ea-9704-e52bdcfdee85", "address": "fa:16:3e:c4:e1:f9", "network": {"id": "6a94bfbb-d0bc-4a6c-9761-901890b34483", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1620466850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c3057c923f84bc1a2b00e9526736e1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3117b312-701b-4439-b197-96b6c5cdca89", "external-id": "nsx-vlan-transportzone-94", "segmentation_id": 94, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0fc4290-34", "ovs_interfaceid": "f0fc4290-34ae-42ea-9704-e52bdcfdee85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.448437] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52cbff4d-1d44-098f-5a33-1fd070cfe9d3, 'name': SearchDatastore_Task, 'duration_secs': 0.019086} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.448437] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.448437] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b/8ff4ebb9-57ea-49d2-8d3c-5876da61e77b.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 713.448437] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a06ebccd-a23a-4750-82a7-93e17307020c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.456260] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Waiting for the task: (returnval){ [ 713.456260] env[61987]: value = "task-1061365" [ 713.456260] env[61987]: _type = "Task" [ 713.456260] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.457797] env[61987]: INFO nova.scheduler.client.report [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Deleted allocations for instance 89a97f81-bcd9-49e5-bbd3-9140f6638a20 [ 713.476552] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': task-1061365, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.482830] env[61987]: DEBUG oslo_vmware.api [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061362, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.508467] env[61987]: INFO nova.compute.manager [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Took 46.67 seconds to build instance. [ 713.605456] env[61987]: DEBUG oslo_vmware.rw_handles [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52984d26-9de9-3da9-53cb-5a3fdc3d9cc4/disk-0.vmdk. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 713.605606] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cdef905-c7ce-4afc-b93b-345a3a36de47 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.613932] env[61987]: DEBUG oslo_vmware.rw_handles [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52984d26-9de9-3da9-53cb-5a3fdc3d9cc4/disk-0.vmdk is in state: ready. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 713.614944] env[61987]: ERROR oslo_vmware.rw_handles [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52984d26-9de9-3da9-53cb-5a3fdc3d9cc4/disk-0.vmdk due to incomplete transfer. [ 713.614944] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-de8978ee-a29a-4cb6-8378-1ec78a8a6ffd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.623594] env[61987]: DEBUG oslo_vmware.rw_handles [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52984d26-9de9-3da9-53cb-5a3fdc3d9cc4/disk-0.vmdk. {{(pid=61987) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 713.623845] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Uploaded image d25beb0f-8ad6-4e15-bf01-99793b52926f to the Glance image server {{(pid=61987) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 713.626422] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Destroying the VM {{(pid=61987) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 713.626729] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6bc28cb5-f75a-450b-ac1c-472ec72b9bc3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.635427] env[61987]: DEBUG oslo_vmware.api [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for the task: (returnval){ [ 713.635427] env[61987]: value = "task-1061366" [ 713.635427] env[61987]: _type = "Task" [ 713.635427] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.644974] env[61987]: DEBUG oslo_vmware.api [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061366, 'name': Destroy_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.653059] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': task-1061363, 'name': ReconfigVM_Task, 'duration_secs': 0.476089} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.653402] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Reconfigured VM instance instance-00000028 to attach disk [datastore2] 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f/49e0d752-6f88-45c4-bd6a-d9fe4f526a2f.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 713.654145] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1b7f5121-e132-4e62-9ae5-5da98571ea3b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.660165] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Waiting for the task: (returnval){ [ 713.660165] env[61987]: value = "task-1061367" [ 713.660165] env[61987]: _type = "Task" [ 713.660165] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.669085] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': task-1061367, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.845444] env[61987]: DEBUG oslo_vmware.api [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061364, 'name': CreateSnapshot_Task, 'duration_secs': 0.495089} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.845896] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Created Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 713.847068] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-596884a2-2c1a-4a14-b92d-8294fc9e9c92 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.942527] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Releasing lock "refresh_cache-e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.942914] env[61987]: DEBUG nova.compute.manager [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Instance network_info: |[{"id": "f0fc4290-34ae-42ea-9704-e52bdcfdee85", "address": "fa:16:3e:c4:e1:f9", "network": {"id": "6a94bfbb-d0bc-4a6c-9761-901890b34483", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1620466850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c3057c923f84bc1a2b00e9526736e1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3117b312-701b-4439-b197-96b6c5cdca89", "external-id": "nsx-vlan-transportzone-94", "segmentation_id": 94, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0fc4290-34", "ovs_interfaceid": "f0fc4290-34ae-42ea-9704-e52bdcfdee85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 713.943357] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:e1:f9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3117b312-701b-4439-b197-96b6c5cdca89', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f0fc4290-34ae-42ea-9704-e52bdcfdee85', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 713.953831] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Creating folder: Project (8c3057c923f84bc1a2b00e9526736e1d). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 713.955916] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-383b8dce-3016-47f3-a118-0121462c0e5c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.976465] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Created folder: Project (8c3057c923f84bc1a2b00e9526736e1d) in parent group-v234219. [ 713.976927] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Creating folder: Instances. Parent ref: group-v234266. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 713.977610] env[61987]: DEBUG oslo_concurrency.lockutils [None req-671d6608-2fc9-4707-8701-eeaf944886f4 tempest-ListImageFiltersTestJSON-1421217924 tempest-ListImageFiltersTestJSON-1421217924-project-member] Lock "89a97f81-bcd9-49e5-bbd3-9140f6638a20" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.144s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.979732] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-83605d11-89f1-40a8-9448-b530fb15308f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.991613] env[61987]: DEBUG oslo_vmware.api [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061362, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.991613] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': task-1061365, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.993379] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Created folder: Instances in parent group-v234266. [ 713.993636] env[61987]: DEBUG oslo.service.loopingcall [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 713.994050] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 713.994342] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-951dcbc2-0bd7-4262-8b9a-8e9542e75241 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.010690] env[61987]: DEBUG oslo_concurrency.lockutils [None req-032d196d-09bb-472b-b952-bbcacac08f48 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Lock "fa943137-caf1-4b41-b644-c9146f72f956" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.712s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.018510] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 714.018510] env[61987]: value = "task-1061370" [ 714.018510] env[61987]: _type = "Task" [ 714.018510] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.025449] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061370, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.145448] env[61987]: DEBUG oslo_vmware.api [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061366, 'name': Destroy_Task} progress is 33%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.170337] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': task-1061367, 'name': Rename_Task, 'duration_secs': 0.268005} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.171158] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 714.171158] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1c58941c-5075-42ee-9d3d-03c4b3358fbf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.177242] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Waiting for the task: (returnval){ [ 714.177242] env[61987]: value = "task-1061371" [ 714.177242] env[61987]: _type = "Task" [ 714.177242] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.186885] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': task-1061371, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.308921] env[61987]: DEBUG nova.network.neutron [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Successfully updated port: fc1b4b12-b396-40ac-9eab-a1309f2259e4 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 714.369519] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Creating linked-clone VM from snapshot {{(pid=61987) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 714.369519] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c6bc1c2a-9091-4296-a8f3-0c0375671a61 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.378302] env[61987]: DEBUG oslo_vmware.api [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 714.378302] env[61987]: value = "task-1061372" [ 714.378302] env[61987]: _type = "Task" [ 714.378302] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.386340] env[61987]: DEBUG oslo_vmware.api [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061372, 'name': CloneVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.465158] env[61987]: DEBUG oslo_concurrency.lockutils [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Acquiring lock "0b4b3bdb-7393-47ab-abc1-44c47496cb3a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.465720] env[61987]: DEBUG oslo_concurrency.lockutils [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Lock "0b4b3bdb-7393-47ab-abc1-44c47496cb3a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.466421] env[61987]: DEBUG oslo_concurrency.lockutils [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Acquiring lock "0b4b3bdb-7393-47ab-abc1-44c47496cb3a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.466421] env[61987]: DEBUG oslo_concurrency.lockutils [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Lock "0b4b3bdb-7393-47ab-abc1-44c47496cb3a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.466421] env[61987]: DEBUG oslo_concurrency.lockutils [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Lock "0b4b3bdb-7393-47ab-abc1-44c47496cb3a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.475898] env[61987]: INFO nova.compute.manager [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Terminating instance [ 714.488880] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': task-1061365, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.585257} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.488962] env[61987]: DEBUG oslo_vmware.api [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061362, 'name': CreateSnapshot_Task, 'duration_secs': 1.049262} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.489193] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b/8ff4ebb9-57ea-49d2-8d3c-5876da61e77b.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 714.489413] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 714.489729] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Created Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 714.490108] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c1420596-d33f-4aad-aac1-81c20bf7acee {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.492715] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6ef35b6-590c-423d-b926-6d13e2afd464 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.509258] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Waiting for the task: (returnval){ [ 714.509258] env[61987]: value = "task-1061373" [ 714.509258] env[61987]: _type = "Task" [ 714.509258] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.518391] env[61987]: DEBUG nova.compute.manager [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 714.528586] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': task-1061373, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.538601] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061370, 'name': CreateVM_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.646222] env[61987]: DEBUG oslo_vmware.api [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061366, 'name': Destroy_Task, 'duration_secs': 0.958261} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.646497] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Destroyed the VM [ 714.646824] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Deleting Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 714.647474] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-f2bfd304-4650-4af7-9fbf-52f4e9fd4a69 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.658340] env[61987]: DEBUG oslo_vmware.api [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for the task: (returnval){ [ 714.658340] env[61987]: value = "task-1061374" [ 714.658340] env[61987]: _type = "Task" [ 714.658340] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.668681] env[61987]: DEBUG oslo_vmware.api [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061374, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.672641] env[61987]: DEBUG nova.compute.manager [None req-f390c623-d4c3-4284-90d4-bc3ce198691a tempest-ServerDiagnosticsTest-554753924 tempest-ServerDiagnosticsTest-554753924-project-admin] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 714.673927] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d7790e7-019b-4084-b0e7-d8e4149ca920 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.687089] env[61987]: INFO nova.compute.manager [None req-f390c623-d4c3-4284-90d4-bc3ce198691a tempest-ServerDiagnosticsTest-554753924 tempest-ServerDiagnosticsTest-554753924-project-admin] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Retrieving diagnostics [ 714.688539] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e058b994-e13d-4250-ac72-49bbf1259184 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.698958] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': task-1061371, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.811646] env[61987]: DEBUG oslo_concurrency.lockutils [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquiring lock "refresh_cache-80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.811799] env[61987]: DEBUG oslo_concurrency.lockutils [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquired lock "refresh_cache-80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.812051] env[61987]: DEBUG nova.network.neutron [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 714.826349] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28ffa4a0-ed09-4d63-9f3d-6ba37305db3d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.834904] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ddd7681-f26f-4abb-9e39-09eef733cb56 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.866650] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a14d7cb4-002a-46b9-9b76-d8f80b82a729 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.874514] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-558711de-6217-49fc-92e2-6e9d0819064f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.890671] env[61987]: DEBUG nova.compute.provider_tree [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.895255] env[61987]: DEBUG nova.compute.manager [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Received event network-changed-f0fc4290-34ae-42ea-9704-e52bdcfdee85 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 714.895255] env[61987]: DEBUG nova.compute.manager [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Refreshing instance network info cache due to event network-changed-f0fc4290-34ae-42ea-9704-e52bdcfdee85. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 714.895255] env[61987]: DEBUG oslo_concurrency.lockutils [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] Acquiring lock "refresh_cache-e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.895255] env[61987]: DEBUG oslo_concurrency.lockutils [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] Acquired lock "refresh_cache-e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.895255] env[61987]: DEBUG nova.network.neutron [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Refreshing network info cache for port f0fc4290-34ae-42ea-9704-e52bdcfdee85 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 714.903612] env[61987]: DEBUG oslo_vmware.api [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061372, 'name': CloneVM_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.984051] env[61987]: DEBUG nova.compute.manager [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 714.984051] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 714.984787] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c5e29f-39be-4a41-a0c0-cea73d5be8a9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.995234] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 714.995575] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c838cdfc-59c5-46d4-aad1-728b2d040fdc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.002581] env[61987]: DEBUG oslo_vmware.api [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Waiting for the task: (returnval){ [ 715.002581] env[61987]: value = "task-1061375" [ 715.002581] env[61987]: _type = "Task" [ 715.002581] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.022171] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Creating linked-clone VM from snapshot {{(pid=61987) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 715.022644] env[61987]: DEBUG oslo_vmware.api [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': task-1061375, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.022970] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6b789931-a821-4a76-9a84-77651ac4cc2c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.047306] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': task-1061373, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068405} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.047698] env[61987]: DEBUG oslo_vmware.api [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 715.047698] env[61987]: value = "task-1061376" [ 715.047698] env[61987]: _type = "Task" [ 715.047698] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.051968] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 715.052401] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061370, 'name': CreateVM_Task, 'duration_secs': 0.735723} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.053382] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f17159c5-c0a9-440d-8d98-775e3349b002 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.060845] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 715.062177] env[61987]: DEBUG oslo_concurrency.lockutils [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.063447] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.063670] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.064098] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 715.065206] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e39eed34-dfe1-4187-a621-e98250fd00ab {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.098298] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Reconfiguring VM instance instance-00000029 to attach disk [datastore2] 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b/8ff4ebb9-57ea-49d2-8d3c-5876da61e77b.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 715.098755] env[61987]: DEBUG oslo_vmware.api [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061376, 'name': CloneVM_Task} progress is 11%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.099604] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9de5f08f-6086-439e-9157-5d4844c1daf5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.123983] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Waiting for the task: (returnval){ [ 715.123983] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]528d19cb-60b1-8316-a0b2-626c3e79cedc" [ 715.123983] env[61987]: _type = "Task" [ 715.123983] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.130373] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Waiting for the task: (returnval){ [ 715.130373] env[61987]: value = "task-1061377" [ 715.130373] env[61987]: _type = "Task" [ 715.130373] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.140304] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]528d19cb-60b1-8316-a0b2-626c3e79cedc, 'name': SearchDatastore_Task, 'duration_secs': 0.013602} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.141234] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.141559] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 715.141862] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.142090] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.142343] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 715.146648] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8a4484e6-714e-4f58-8eee-9ffc6d141c96 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.148921] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': task-1061377, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.155974] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 715.156232] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 715.157022] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc34d192-ab90-4326-b871-15ef2d02de30 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.166353] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Waiting for the task: (returnval){ [ 715.166353] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52cc7c79-a686-74e3-a62a-ec49dbc6f26b" [ 715.166353] env[61987]: _type = "Task" [ 715.166353] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.169536] env[61987]: DEBUG oslo_vmware.api [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061374, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.179155] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52cc7c79-a686-74e3-a62a-ec49dbc6f26b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.191947] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': task-1061371, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.364088] env[61987]: DEBUG nova.network.neutron [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 715.391301] env[61987]: DEBUG oslo_vmware.api [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061372, 'name': CloneVM_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.399476] env[61987]: DEBUG nova.scheduler.client.report [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 715.519093] env[61987]: DEBUG oslo_vmware.api [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': task-1061375, 'name': PowerOffVM_Task, 'duration_secs': 0.32171} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.519464] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 715.519667] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 715.520018] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ad098e77-b5dc-4cd5-82b7-34e9f299db47 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.566837] env[61987]: DEBUG oslo_vmware.api [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061376, 'name': CloneVM_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.642468] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': task-1061377, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.670170] env[61987]: DEBUG oslo_vmware.api [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061374, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.680180] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52cc7c79-a686-74e3-a62a-ec49dbc6f26b, 'name': SearchDatastore_Task, 'duration_secs': 0.018749} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.684766] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45679e7c-4b31-4d5e-99bc-65c9d8ad1db9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.693776] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Waiting for the task: (returnval){ [ 715.693776] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5272f5c4-f228-4166-4dbd-33bbb4c4d4da" [ 715.693776] env[61987]: _type = "Task" [ 715.693776] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.697027] env[61987]: DEBUG oslo_vmware.api [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': task-1061371, 'name': PowerOnVM_Task, 'duration_secs': 1.213318} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.700189] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 715.700411] env[61987]: INFO nova.compute.manager [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Took 9.64 seconds to spawn the instance on the hypervisor. [ 715.700591] env[61987]: DEBUG nova.compute.manager [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 715.701715] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a681acba-68d8-4d5f-8241-d2c55836a962 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.709033] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5272f5c4-f228-4166-4dbd-33bbb4c4d4da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.873661] env[61987]: DEBUG nova.network.neutron [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Updating instance_info_cache with network_info: [{"id": "fc1b4b12-b396-40ac-9eab-a1309f2259e4", "address": "fa:16:3e:ec:35:ab", "network": {"id": "8a9770a4-fed3-4232-ac78-8a1fc2223229", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.118", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "705320a8b6a549d0b197ee1a35404d52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bafe8721-91d4-4127-b215-d9e8e27947dc", "external-id": "nsx-vlan-transportzone-680", "segmentation_id": 680, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc1b4b12-b3", "ovs_interfaceid": "fc1b4b12-b396-40ac-9eab-a1309f2259e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.891013] env[61987]: DEBUG oslo_vmware.api [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061372, 'name': CloneVM_Task} progress is 95%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.909078] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.533s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.909078] env[61987]: DEBUG nova.compute.manager [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 715.910741] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.499s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.912158] env[61987]: INFO nova.compute.claims [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 716.007601] env[61987]: DEBUG nova.network.neutron [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Updated VIF entry in instance network info cache for port f0fc4290-34ae-42ea-9704-e52bdcfdee85. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 716.008074] env[61987]: DEBUG nova.network.neutron [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Updating instance_info_cache with network_info: [{"id": "f0fc4290-34ae-42ea-9704-e52bdcfdee85", "address": "fa:16:3e:c4:e1:f9", "network": {"id": "6a94bfbb-d0bc-4a6c-9761-901890b34483", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1620466850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c3057c923f84bc1a2b00e9526736e1d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3117b312-701b-4439-b197-96b6c5cdca89", "external-id": "nsx-vlan-transportzone-94", "segmentation_id": 94, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0fc4290-34", "ovs_interfaceid": "f0fc4290-34ae-42ea-9704-e52bdcfdee85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.063242] env[61987]: DEBUG oslo_vmware.api [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061376, 'name': CloneVM_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.142384] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': task-1061377, 'name': ReconfigVM_Task, 'duration_secs': 0.548916} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.142684] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Reconfigured VM instance instance-00000029 to attach disk [datastore2] 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b/8ff4ebb9-57ea-49d2-8d3c-5876da61e77b.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 716.143361] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-36c7051d-99e0-4a22-a9b3-70705d0a23ac {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.150016] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Waiting for the task: (returnval){ [ 716.150016] env[61987]: value = "task-1061379" [ 716.150016] env[61987]: _type = "Task" [ 716.150016] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.160034] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': task-1061379, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.168974] env[61987]: DEBUG oslo_vmware.api [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061374, 'name': RemoveSnapshot_Task, 'duration_secs': 1.12918} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.169290] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Deleted Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 716.169537] env[61987]: INFO nova.compute.manager [None req-e76d1e86-c2dd-40df-9f3b-7701d7baa93b tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Took 18.56 seconds to snapshot the instance on the hypervisor. [ 716.209210] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5272f5c4-f228-4166-4dbd-33bbb4c4d4da, 'name': SearchDatastore_Task, 'duration_secs': 0.038079} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.209210] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.209210] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1/e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 716.209210] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-78bbb955-6d52-4043-8d7c-6a931b7b389c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.218422] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Waiting for the task: (returnval){ [ 716.218422] env[61987]: value = "task-1061380" [ 716.218422] env[61987]: _type = "Task" [ 716.218422] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.223880] env[61987]: INFO nova.compute.manager [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Took 47.40 seconds to build instance. [ 716.229734] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': task-1061380, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.262051] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Acquiring lock "fa943137-caf1-4b41-b644-c9146f72f956" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.262349] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Lock "fa943137-caf1-4b41-b644-c9146f72f956" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.262559] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Acquiring lock "fa943137-caf1-4b41-b644-c9146f72f956-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.262752] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Lock "fa943137-caf1-4b41-b644-c9146f72f956-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.262931] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Lock "fa943137-caf1-4b41-b644-c9146f72f956-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.265296] env[61987]: INFO nova.compute.manager [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Terminating instance [ 716.376781] env[61987]: DEBUG oslo_concurrency.lockutils [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Releasing lock "refresh_cache-80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.377017] env[61987]: DEBUG nova.compute.manager [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Instance network_info: |[{"id": "fc1b4b12-b396-40ac-9eab-a1309f2259e4", "address": "fa:16:3e:ec:35:ab", "network": {"id": "8a9770a4-fed3-4232-ac78-8a1fc2223229", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.118", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "705320a8b6a549d0b197ee1a35404d52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bafe8721-91d4-4127-b215-d9e8e27947dc", "external-id": "nsx-vlan-transportzone-680", "segmentation_id": 680, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc1b4b12-b3", "ovs_interfaceid": "fc1b4b12-b396-40ac-9eab-a1309f2259e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 716.377478] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:35:ab', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bafe8721-91d4-4127-b215-d9e8e27947dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fc1b4b12-b396-40ac-9eab-a1309f2259e4', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 716.385204] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Creating folder: Project (d4e086581d9e41b197f84dac49aab72e). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 716.385492] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5002c9e7-a7ac-4cb8-a767-79eec1de1a2f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.396116] env[61987]: DEBUG oslo_vmware.api [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061372, 'name': CloneVM_Task} progress is 95%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.398062] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Created folder: Project (d4e086581d9e41b197f84dac49aab72e) in parent group-v234219. [ 716.398265] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Creating folder: Instances. Parent ref: group-v234272. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 716.398494] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bcfbabb0-299f-4a80-be68-70129c58523c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.406578] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Created folder: Instances in parent group-v234272. [ 716.406865] env[61987]: DEBUG oslo.service.loopingcall [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 716.407014] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 716.407225] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e00dec61-3da4-4932-9d0f-f9bf19600654 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.423089] env[61987]: DEBUG nova.compute.utils [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 716.426876] env[61987]: DEBUG nova.compute.manager [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 716.426876] env[61987]: DEBUG nova.network.neutron [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 716.435398] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 716.435398] env[61987]: value = "task-1061383" [ 716.435398] env[61987]: _type = "Task" [ 716.435398] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.444296] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061383, 'name': CreateVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.514905] env[61987]: DEBUG oslo_concurrency.lockutils [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] Releasing lock "refresh_cache-e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.515360] env[61987]: DEBUG nova.compute.manager [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Received event network-vif-plugged-fc1b4b12-b396-40ac-9eab-a1309f2259e4 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 716.515707] env[61987]: DEBUG oslo_concurrency.lockutils [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] Acquiring lock "80f13fca-2df1-4aa0-96f1-b61a60c4fdc6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.516098] env[61987]: DEBUG oslo_concurrency.lockutils [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] Lock "80f13fca-2df1-4aa0-96f1-b61a60c4fdc6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.516402] env[61987]: DEBUG oslo_concurrency.lockutils [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] Lock "80f13fca-2df1-4aa0-96f1-b61a60c4fdc6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.516728] env[61987]: DEBUG nova.compute.manager [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] No waiting events found dispatching network-vif-plugged-fc1b4b12-b396-40ac-9eab-a1309f2259e4 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 716.517302] env[61987]: WARNING nova.compute.manager [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Received unexpected event network-vif-plugged-fc1b4b12-b396-40ac-9eab-a1309f2259e4 for instance with vm_state building and task_state spawning. [ 716.517716] env[61987]: DEBUG nova.compute.manager [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Received event network-changed-fc1b4b12-b396-40ac-9eab-a1309f2259e4 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 716.518037] env[61987]: DEBUG nova.compute.manager [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Refreshing instance network info cache due to event network-changed-fc1b4b12-b396-40ac-9eab-a1309f2259e4. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 716.518462] env[61987]: DEBUG oslo_concurrency.lockutils [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] Acquiring lock "refresh_cache-80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.518629] env[61987]: DEBUG oslo_concurrency.lockutils [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] Acquired lock "refresh_cache-80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.518885] env[61987]: DEBUG nova.network.neutron [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Refreshing network info cache for port fc1b4b12-b396-40ac-9eab-a1309f2259e4 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 716.522444] env[61987]: DEBUG nova.policy [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '15152ee0aa1f4d9db7e63ab08ad76a4b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2287d9f275945a5959acb3e6ef26074', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 716.527497] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 716.527951] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 716.528319] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Deleting the datastore file [datastore1] 0b4b3bdb-7393-47ab-abc1-44c47496cb3a {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 716.529420] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2b554adc-cf6d-4380-bc4f-9579d5c24623 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.537611] env[61987]: DEBUG oslo_vmware.api [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Waiting for the task: (returnval){ [ 716.537611] env[61987]: value = "task-1061384" [ 716.537611] env[61987]: _type = "Task" [ 716.537611] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.549150] env[61987]: DEBUG oslo_vmware.api [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': task-1061384, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.562497] env[61987]: DEBUG oslo_vmware.api [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061376, 'name': CloneVM_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.659757] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': task-1061379, 'name': Rename_Task, 'duration_secs': 0.405796} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.660064] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 716.660415] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f6a50646-3e51-4002-bd13-6049dda0ae28 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.665605] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Waiting for the task: (returnval){ [ 716.665605] env[61987]: value = "task-1061385" [ 716.665605] env[61987]: _type = "Task" [ 716.665605] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.675663] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': task-1061385, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.726644] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ce230cea-593b-4f17-abaa-9f575a26d36b tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Lock "49e0d752-6f88-45c4-bd6a-d9fe4f526a2f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.814s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.733614] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': task-1061380, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.770320] env[61987]: DEBUG nova.compute.manager [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 716.770722] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 716.771963] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe17319a-d6db-4cba-a835-19e143bffbc9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.782119] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 716.782301] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ad5e243-43cc-4898-adfe-7b2807a07b6e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.790333] env[61987]: DEBUG oslo_vmware.api [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Waiting for the task: (returnval){ [ 716.790333] env[61987]: value = "task-1061386" [ 716.790333] env[61987]: _type = "Task" [ 716.790333] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.802324] env[61987]: DEBUG oslo_vmware.api [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': task-1061386, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.896389] env[61987]: DEBUG oslo_vmware.api [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061372, 'name': CloneVM_Task, 'duration_secs': 2.237875} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.896555] env[61987]: INFO nova.virt.vmwareapi.vmops [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Created linked-clone VM from snapshot [ 716.897276] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07ec31c3-99c3-4513-93e1-c58ac0f843e0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.905600] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Uploading image d76b7594-8ff0-479d-94b4-4cd677bf22ab {{(pid=61987) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 716.925628] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Destroying the VM {{(pid=61987) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 716.928033] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-eb631c98-2446-4039-b456-f5b3287545d1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.933283] env[61987]: DEBUG nova.compute.manager [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 716.940912] env[61987]: DEBUG oslo_vmware.api [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 716.940912] env[61987]: value = "task-1061387" [ 716.940912] env[61987]: _type = "Task" [ 716.940912] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.950800] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061383, 'name': CreateVM_Task, 'duration_secs': 0.435732} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.951658] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 716.952468] env[61987]: DEBUG oslo_concurrency.lockutils [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.952703] env[61987]: DEBUG oslo_concurrency.lockutils [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.953103] env[61987]: DEBUG oslo_concurrency.lockutils [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 716.953431] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c3e8855-de69-4f58-93bc-53d43e6cf002 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.961562] env[61987]: DEBUG oslo_vmware.api [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061387, 'name': Destroy_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.963371] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for the task: (returnval){ [ 716.963371] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]526760ec-07dd-e2d7-5c7b-eab1cd7970d2" [ 716.963371] env[61987]: _type = "Task" [ 716.963371] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.972439] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]526760ec-07dd-e2d7-5c7b-eab1cd7970d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.049145] env[61987]: DEBUG oslo_vmware.api [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Task: {'id': task-1061384, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.22165} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.052201] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 717.052366] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 717.052568] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 717.052775] env[61987]: INFO nova.compute.manager [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Took 2.07 seconds to destroy the instance on the hypervisor. [ 717.053233] env[61987]: DEBUG oslo.service.loopingcall [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 717.053898] env[61987]: DEBUG nova.compute.manager [-] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 717.054069] env[61987]: DEBUG nova.network.neutron [-] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 717.068385] env[61987]: DEBUG oslo_vmware.api [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061376, 'name': CloneVM_Task, 'duration_secs': 1.763751} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.068636] env[61987]: INFO nova.virt.vmwareapi.vmops [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Created linked-clone VM from snapshot [ 717.069722] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c76e4fab-5745-47e5-9b0c-32bc14b75066 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.081682] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Uploading image f31f91a0-19e6-447b-abcc-cd75f8cffd89 {{(pid=61987) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 717.113586] env[61987]: DEBUG oslo_vmware.rw_handles [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 717.113586] env[61987]: value = "vm-234271" [ 717.113586] env[61987]: _type = "VirtualMachine" [ 717.113586] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 717.113937] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-54237f17-f7e6-45aa-b11e-f0c9bfb41fa8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.121942] env[61987]: DEBUG oslo_vmware.rw_handles [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lease: (returnval){ [ 717.121942] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52f49919-6dbb-6a94-40b5-f90c24bc86c5" [ 717.121942] env[61987]: _type = "HttpNfcLease" [ 717.121942] env[61987]: } obtained for exporting VM: (result){ [ 717.121942] env[61987]: value = "vm-234271" [ 717.121942] env[61987]: _type = "VirtualMachine" [ 717.121942] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 717.122331] env[61987]: DEBUG oslo_vmware.api [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the lease: (returnval){ [ 717.122331] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52f49919-6dbb-6a94-40b5-f90c24bc86c5" [ 717.122331] env[61987]: _type = "HttpNfcLease" [ 717.122331] env[61987]: } to be ready. {{(pid=61987) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 717.130782] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 717.130782] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52f49919-6dbb-6a94-40b5-f90c24bc86c5" [ 717.130782] env[61987]: _type = "HttpNfcLease" [ 717.130782] env[61987]: } is initializing. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 717.151514] env[61987]: DEBUG nova.network.neutron [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Successfully created port: b3fafafe-77c6-41b9-ae56-b17b2d781bca {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 717.182536] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': task-1061385, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.232264] env[61987]: DEBUG nova.compute.manager [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 717.237881] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': task-1061380, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.307641] env[61987]: DEBUG oslo_vmware.api [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': task-1061386, 'name': PowerOffVM_Task, 'duration_secs': 0.442683} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.308951] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 717.309924] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 717.310286] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f4788b41-5ece-49c7-bf8a-ab3e0e4f29d4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.361470] env[61987]: DEBUG oslo_concurrency.lockutils [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Acquiring lock "49e0d752-6f88-45c4-bd6a-d9fe4f526a2f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.361889] env[61987]: DEBUG oslo_concurrency.lockutils [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Lock "49e0d752-6f88-45c4-bd6a-d9fe4f526a2f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.362152] env[61987]: DEBUG oslo_concurrency.lockutils [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Acquiring lock "49e0d752-6f88-45c4-bd6a-d9fe4f526a2f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.362381] env[61987]: DEBUG oslo_concurrency.lockutils [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Lock "49e0d752-6f88-45c4-bd6a-d9fe4f526a2f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.362564] env[61987]: DEBUG oslo_concurrency.lockutils [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Lock "49e0d752-6f88-45c4-bd6a-d9fe4f526a2f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.367129] env[61987]: INFO nova.compute.manager [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Terminating instance [ 717.408254] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 717.408520] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 717.408686] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Deleting the datastore file [datastore2] fa943137-caf1-4b41-b644-c9146f72f956 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 717.411941] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-43107b1c-2908-482e-916c-745406d59e90 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.416431] env[61987]: DEBUG oslo_vmware.api [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Waiting for the task: (returnval){ [ 717.416431] env[61987]: value = "task-1061390" [ 717.416431] env[61987]: _type = "Task" [ 717.416431] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.430488] env[61987]: DEBUG oslo_vmware.api [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': task-1061390, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.455989] env[61987]: DEBUG oslo_vmware.api [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061387, 'name': Destroy_Task, 'duration_secs': 0.377513} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.456296] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Destroyed the VM [ 717.456537] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Deleting Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 717.456795] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-60b12f50-1448-465d-afaa-aa80053d1447 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.465345] env[61987]: DEBUG oslo_vmware.api [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 717.465345] env[61987]: value = "task-1061391" [ 717.465345] env[61987]: _type = "Task" [ 717.465345] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.482016] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b768dd4e-0e60-4180-8d04-add4f9880411 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.489596] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]526760ec-07dd-e2d7-5c7b-eab1cd7970d2, 'name': SearchDatastore_Task, 'duration_secs': 0.013261} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.490128] env[61987]: DEBUG oslo_concurrency.lockutils [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.490456] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 717.490715] env[61987]: DEBUG oslo_concurrency.lockutils [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.490862] env[61987]: DEBUG oslo_concurrency.lockutils [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.491321] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 717.496499] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6acec4e6-f9bf-4185-834a-0cda469cb71f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.498412] env[61987]: DEBUG oslo_vmware.api [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061391, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.499702] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baefb54a-9aab-4742-891d-0de2b294dcdc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.539052] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe557287-e44d-464a-ac79-79155ede8fb2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.542107] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 717.542307] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 717.544595] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-562c2352-0ca6-4dec-81ae-2b95bc617cca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.556523] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83412887-57fb-4269-bc0c-422c39ca1801 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.560590] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for the task: (returnval){ [ 717.560590] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52a1d4a5-bba9-d1c7-c247-cf5c1bfa0c59" [ 717.560590] env[61987]: _type = "Task" [ 717.560590] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.578444] env[61987]: DEBUG nova.compute.provider_tree [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.584141] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52a1d4a5-bba9-d1c7-c247-cf5c1bfa0c59, 'name': SearchDatastore_Task, 'duration_secs': 0.026582} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.585426] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e49c245-d0f7-4c78-b81c-9d88917b743a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.590545] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for the task: (returnval){ [ 717.590545] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]522daad6-1b71-52ae-0f4f-36be3c9719d8" [ 717.590545] env[61987]: _type = "Task" [ 717.590545] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.599136] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]522daad6-1b71-52ae-0f4f-36be3c9719d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.631376] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 717.631376] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52f49919-6dbb-6a94-40b5-f90c24bc86c5" [ 717.631376] env[61987]: _type = "HttpNfcLease" [ 717.631376] env[61987]: } is ready. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 717.631690] env[61987]: DEBUG oslo_vmware.rw_handles [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 717.631690] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52f49919-6dbb-6a94-40b5-f90c24bc86c5" [ 717.631690] env[61987]: _type = "HttpNfcLease" [ 717.631690] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 717.632496] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afb6b3cb-1987-452f-8e48-40a90b0267d3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.640466] env[61987]: DEBUG oslo_vmware.rw_handles [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5253a660-d193-da86-08f5-d70d89a16a13/disk-0.vmdk from lease info. {{(pid=61987) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 717.640689] env[61987]: DEBUG oslo_vmware.rw_handles [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5253a660-d193-da86-08f5-d70d89a16a13/disk-0.vmdk for reading. {{(pid=61987) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 717.714875] env[61987]: DEBUG oslo_vmware.api [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': task-1061385, 'name': PowerOnVM_Task, 'duration_secs': 1.017767} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.715498] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 717.715843] env[61987]: INFO nova.compute.manager [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Took 10.18 seconds to spawn the instance on the hypervisor. [ 717.716161] env[61987]: DEBUG nova.compute.manager [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 717.717458] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd92e6f4-af11-441a-97c3-d6c71ef6cfd1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.734320] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': task-1061380, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.237826} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.734632] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1/e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 717.734813] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 717.736018] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c2718d01-f366-4640-adf5-99f5e2bac2f0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.742368] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Waiting for the task: (returnval){ [ 717.742368] env[61987]: value = "task-1061392" [ 717.742368] env[61987]: _type = "Task" [ 717.742368] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.749234] env[61987]: DEBUG nova.network.neutron [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Updated VIF entry in instance network info cache for port fc1b4b12-b396-40ac-9eab-a1309f2259e4. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 717.749918] env[61987]: DEBUG nova.network.neutron [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Updating instance_info_cache with network_info: [{"id": "fc1b4b12-b396-40ac-9eab-a1309f2259e4", "address": "fa:16:3e:ec:35:ab", "network": {"id": "8a9770a4-fed3-4232-ac78-8a1fc2223229", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.118", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "705320a8b6a549d0b197ee1a35404d52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bafe8721-91d4-4127-b215-d9e8e27947dc", "external-id": "nsx-vlan-transportzone-680", "segmentation_id": 680, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc1b4b12-b3", "ovs_interfaceid": "fc1b4b12-b396-40ac-9eab-a1309f2259e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.757835] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': task-1061392, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.774909] env[61987]: DEBUG oslo_concurrency.lockutils [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.778303] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c31590e6-e419-4cbf-80f8-bc3cc05cac63 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.871290] env[61987]: DEBUG nova.compute.manager [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 717.874016] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 717.874016] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15da8b4a-b2c6-4693-90e1-779f4b4bef2e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.881334] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 717.881787] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-812463a3-a5af-4a41-a18e-406822a9f982 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.887979] env[61987]: DEBUG oslo_vmware.api [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Waiting for the task: (returnval){ [ 717.887979] env[61987]: value = "task-1061393" [ 717.887979] env[61987]: _type = "Task" [ 717.887979] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.899012] env[61987]: DEBUG oslo_vmware.api [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': task-1061393, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.926264] env[61987]: DEBUG oslo_vmware.api [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': task-1061390, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.943142] env[61987]: DEBUG nova.compute.manager [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 717.962862] env[61987]: DEBUG nova.network.neutron [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Successfully created port: a4965cbd-fb11-4afb-acec-bf0128f32e0f {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 717.973152] env[61987]: DEBUG nova.virt.hardware [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 717.973152] env[61987]: DEBUG nova.virt.hardware [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 717.973152] env[61987]: DEBUG nova.virt.hardware [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 717.973152] env[61987]: DEBUG nova.virt.hardware [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 717.973333] env[61987]: DEBUG nova.virt.hardware [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 717.973333] env[61987]: DEBUG nova.virt.hardware [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 717.973333] env[61987]: DEBUG nova.virt.hardware [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 717.973333] env[61987]: DEBUG nova.virt.hardware [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 717.973333] env[61987]: DEBUG nova.virt.hardware [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 717.973508] env[61987]: DEBUG nova.virt.hardware [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 717.973508] env[61987]: DEBUG nova.virt.hardware [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 717.974244] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcae92e0-8725-4394-a001-9e9ce320a18a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.986228] env[61987]: DEBUG oslo_vmware.api [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061391, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.989590] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da3b9d49-9488-4581-b53e-85083e4c8050 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.081309] env[61987]: DEBUG nova.scheduler.client.report [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 718.101944] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]522daad6-1b71-52ae-0f4f-36be3c9719d8, 'name': SearchDatastore_Task, 'duration_secs': 0.012625} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.102328] env[61987]: DEBUG oslo_concurrency.lockutils [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.102738] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6/80f13fca-2df1-4aa0-96f1-b61a60c4fdc6.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 718.103070] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1e7b05c5-78a5-493f-ba10-b3370c693379 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.109521] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for the task: (returnval){ [ 718.109521] env[61987]: value = "task-1061394" [ 718.109521] env[61987]: _type = "Task" [ 718.109521] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.118724] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061394, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.242873] env[61987]: INFO nova.compute.manager [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Took 33.07 seconds to build instance. [ 718.255098] env[61987]: DEBUG oslo_concurrency.lockutils [req-7130bece-5514-4fbe-b866-f933787141e3 req-bf4dcabd-8d95-4808-984a-0cd3f54d0926 service nova] Releasing lock "refresh_cache-80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.255546] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': task-1061392, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.260237} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.255788] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 718.256665] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c7999be-ff3b-46b2-8ae8-363f9a761f14 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.281966] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Reconfiguring VM instance instance-0000002a to attach disk [datastore2] e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1/e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 718.283603] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5a530550-09c5-4850-aa4c-ed9664815edb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.306595] env[61987]: DEBUG nova.compute.manager [req-1a23ca0a-430b-4a4f-9542-2ea1c66270d6 req-4a675950-982b-48d7-91b7-e5598a57c532 service nova] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Received event network-vif-deleted-2b8e0044-50eb-40c8-8a96-9c828626d2f7 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 718.306841] env[61987]: INFO nova.compute.manager [req-1a23ca0a-430b-4a4f-9542-2ea1c66270d6 req-4a675950-982b-48d7-91b7-e5598a57c532 service nova] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Neutron deleted interface 2b8e0044-50eb-40c8-8a96-9c828626d2f7; detaching it from the instance and deleting it from the info cache [ 718.307057] env[61987]: DEBUG nova.network.neutron [req-1a23ca0a-430b-4a4f-9542-2ea1c66270d6 req-4a675950-982b-48d7-91b7-e5598a57c532 service nova] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.317033] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Waiting for the task: (returnval){ [ 718.317033] env[61987]: value = "task-1061395" [ 718.317033] env[61987]: _type = "Task" [ 718.317033] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.327591] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': task-1061395, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.355444] env[61987]: DEBUG nova.network.neutron [-] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.413039] env[61987]: DEBUG oslo_vmware.api [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': task-1061393, 'name': PowerOffVM_Task, 'duration_secs': 0.487763} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.413039] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 718.413039] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 718.413263] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-53f5ffde-312f-44a5-a19d-a04119d11c35 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.430728] env[61987]: DEBUG oslo_vmware.api [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Task: {'id': task-1061390, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.657932} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.431122] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 718.431553] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 718.431889] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 718.432931] env[61987]: INFO nova.compute.manager [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Took 1.66 seconds to destroy the instance on the hypervisor. [ 718.433329] env[61987]: DEBUG oslo.service.loopingcall [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 718.433648] env[61987]: DEBUG nova.compute.manager [-] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 718.433767] env[61987]: DEBUG nova.network.neutron [-] [instance: fa943137-caf1-4b41-b644-c9146f72f956] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 718.484672] env[61987]: DEBUG oslo_vmware.api [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061391, 'name': RemoveSnapshot_Task, 'duration_secs': 0.837963} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.487877] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Deleted Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 718.586222] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.675s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.587394] env[61987]: DEBUG nova.compute.manager [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 718.591543] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.134s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.592707] env[61987]: INFO nova.compute.claims [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 718.622741] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061394, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474367} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.622741] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6/80f13fca-2df1-4aa0-96f1-b61a60c4fdc6.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 718.622741] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 718.622741] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-29f6a6d8-b21d-4e89-b500-bde347ab4f9f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.628469] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for the task: (returnval){ [ 718.628469] env[61987]: value = "task-1061397" [ 718.628469] env[61987]: _type = "Task" [ 718.628469] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.638677] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061397, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.639320] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 718.639895] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 718.639895] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Deleting the datastore file [datastore2] 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 718.640050] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-28406704-9692-4022-b31a-22874db2da94 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.647712] env[61987]: DEBUG oslo_vmware.api [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Waiting for the task: (returnval){ [ 718.647712] env[61987]: value = "task-1061398" [ 718.647712] env[61987]: _type = "Task" [ 718.647712] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.657144] env[61987]: DEBUG oslo_vmware.api [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': task-1061398, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.745035] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ada078d-c805-40b7-bcf2-e7e40b96751d tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Lock "8ff4ebb9-57ea-49d2-8d3c-5876da61e77b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.242s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.811969] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-747d5bd9-9b4a-47ac-b6be-03cce7ee3031 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.827300] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb903d8a-ebc8-4327-83ab-d85318027717 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.847652] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': task-1061395, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.866753] env[61987]: INFO nova.compute.manager [-] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Took 1.81 seconds to deallocate network for instance. [ 718.867276] env[61987]: DEBUG nova.compute.manager [req-1a23ca0a-430b-4a4f-9542-2ea1c66270d6 req-4a675950-982b-48d7-91b7-e5598a57c532 service nova] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Detach interface failed, port_id=2b8e0044-50eb-40c8-8a96-9c828626d2f7, reason: Instance 0b4b3bdb-7393-47ab-abc1-44c47496cb3a could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 718.998741] env[61987]: WARNING nova.compute.manager [None req-64244713-8255-4293-8464-388607d5521f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Image not found during snapshot: nova.exception.ImageNotFound: Image d76b7594-8ff0-479d-94b4-4cd677bf22ab could not be found. [ 719.096706] env[61987]: DEBUG nova.compute.utils [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 719.100337] env[61987]: DEBUG nova.compute.manager [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 719.100522] env[61987]: DEBUG nova.network.neutron [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 719.140386] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061397, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073787} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.140734] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 719.141624] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e348c456-9b15-48bf-8ad0-22c101d5d82b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.146353] env[61987]: DEBUG nova.policy [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7017f711a0164de5aea988eb5ba1ea65', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8a22d437ed0e42089681884b0d28c182', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 719.173972] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Reconfiguring VM instance instance-0000002b to attach disk [datastore1] 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6/80f13fca-2df1-4aa0-96f1-b61a60c4fdc6.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 719.176739] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd59b1d2-a686-4905-8e80-19cdd22909b0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.196974] env[61987]: DEBUG oslo_vmware.api [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': task-1061398, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.202875] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for the task: (returnval){ [ 719.202875] env[61987]: value = "task-1061399" [ 719.202875] env[61987]: _type = "Task" [ 719.202875] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.212904] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061399, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.247349] env[61987]: DEBUG nova.compute.manager [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 719.328532] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': task-1061395, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.375335] env[61987]: DEBUG oslo_concurrency.lockutils [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.444895] env[61987]: DEBUG nova.network.neutron [-] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.604898] env[61987]: DEBUG nova.compute.manager [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 719.663787] env[61987]: DEBUG oslo_vmware.api [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Task: {'id': task-1061398, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.650255} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.663787] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 719.664164] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 719.664164] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 719.664274] env[61987]: INFO nova.compute.manager [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Took 1.79 seconds to destroy the instance on the hypervisor. [ 719.664515] env[61987]: DEBUG oslo.service.loopingcall [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 719.664754] env[61987]: DEBUG nova.compute.manager [-] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 719.668029] env[61987]: DEBUG nova.network.neutron [-] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 719.668029] env[61987]: DEBUG nova.network.neutron [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Successfully created port: 27b82c05-3eb6-48ab-9455-b321ef1f1fe6 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 719.697214] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "ea6c2120-a69a-42c7-bc92-3e74539c008d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.697526] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "ea6c2120-a69a-42c7-bc92-3e74539c008d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.697764] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "ea6c2120-a69a-42c7-bc92-3e74539c008d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.697961] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "ea6c2120-a69a-42c7-bc92-3e74539c008d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.698179] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "ea6c2120-a69a-42c7-bc92-3e74539c008d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.703659] env[61987]: INFO nova.compute.manager [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Terminating instance [ 719.716693] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061399, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.773373] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.828860] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': task-1061395, 'name': ReconfigVM_Task, 'duration_secs': 1.184164} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.829202] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Reconfigured VM instance instance-0000002a to attach disk [datastore2] e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1/e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 719.832499] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7415befe-86c4-4ed8-9573-c01925ed337d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.839765] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Waiting for the task: (returnval){ [ 719.839765] env[61987]: value = "task-1061400" [ 719.839765] env[61987]: _type = "Task" [ 719.839765] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.848924] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': task-1061400, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.947553] env[61987]: INFO nova.compute.manager [-] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Took 1.51 seconds to deallocate network for instance. [ 720.041304] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c749af-0893-4158-b6b3-81fd37cfcd82 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.049959] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd31d482-a120-4724-9109-3f2eb607e169 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.084773] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f556fb0-edeb-4cfd-afde-56751d106dec {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.092675] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6548790-6000-431e-946b-4b84a9bcab86 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.109176] env[61987]: DEBUG nova.compute.provider_tree [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.211401] env[61987]: DEBUG nova.compute.manager [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 720.211782] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 720.216114] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71560fcf-7da1-49a4-8c22-7cb061bad786 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.219393] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061399, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.224128] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 720.225301] env[61987]: DEBUG nova.network.neutron [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Successfully updated port: b3fafafe-77c6-41b9-ae56-b17b2d781bca {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 720.226453] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-98203bfe-a378-4eaf-96e2-bdbcaeab9c61 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.235134] env[61987]: DEBUG oslo_vmware.api [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 720.235134] env[61987]: value = "task-1061401" [ 720.235134] env[61987]: _type = "Task" [ 720.235134] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.242884] env[61987]: DEBUG oslo_vmware.api [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061401, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.351388] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': task-1061400, 'name': Rename_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.457673] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.614397] env[61987]: DEBUG nova.scheduler.client.report [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 720.619413] env[61987]: DEBUG nova.compute.manager [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 720.623327] env[61987]: DEBUG nova.compute.manager [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 720.624688] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1e7c917-691d-4bb5-b15e-576ab268b1a3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.655256] env[61987]: DEBUG nova.virt.hardware [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 720.655616] env[61987]: DEBUG nova.virt.hardware [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 720.655834] env[61987]: DEBUG nova.virt.hardware [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 720.657120] env[61987]: DEBUG nova.virt.hardware [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 720.657374] env[61987]: DEBUG nova.virt.hardware [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 720.657610] env[61987]: DEBUG nova.virt.hardware [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 720.657873] env[61987]: DEBUG nova.virt.hardware [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 720.658110] env[61987]: DEBUG nova.virt.hardware [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 720.658341] env[61987]: DEBUG nova.virt.hardware [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 720.658571] env[61987]: DEBUG nova.virt.hardware [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 720.658793] env[61987]: DEBUG nova.virt.hardware [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 720.659762] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f68cb8d1-c510-4a37-a8bf-2766ade4fa6f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.668584] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06f1ed77-1e36-461e-80fc-456e788abb84 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.715119] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061399, 'name': ReconfigVM_Task, 'duration_secs': 1.452115} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.715822] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Reconfigured VM instance instance-0000002b to attach disk [datastore1] 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6/80f13fca-2df1-4aa0-96f1-b61a60c4fdc6.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 720.716491] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd5d3cb3-c2b8-4753-ab3f-09bd13c2ef0a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.723154] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for the task: (returnval){ [ 720.723154] env[61987]: value = "task-1061402" [ 720.723154] env[61987]: _type = "Task" [ 720.723154] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.733507] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061402, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.743018] env[61987]: DEBUG oslo_vmware.api [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061401, 'name': PowerOffVM_Task, 'duration_secs': 0.236893} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.743018] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 720.743018] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 720.743018] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4f001b09-75d4-4563-a2d5-822695f7790e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.806378] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 720.806628] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 720.806816] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Deleting the datastore file [datastore1] ea6c2120-a69a-42c7-bc92-3e74539c008d {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 720.807100] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4bdbdef1-e063-47cd-a110-95b6bd58224e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.815502] env[61987]: DEBUG oslo_vmware.api [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 720.815502] env[61987]: value = "task-1061404" [ 720.815502] env[61987]: _type = "Task" [ 720.815502] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.828079] env[61987]: DEBUG oslo_vmware.api [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061404, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.836731] env[61987]: DEBUG nova.network.neutron [-] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.854277] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': task-1061400, 'name': Rename_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.921507] env[61987]: DEBUG nova.compute.manager [req-6f125fcc-e7c9-42de-997a-570902b4250e req-e5cbb9b0-a0a0-4946-b37e-a403db2d0a19 service nova] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Received event network-vif-deleted-5134697b-ec80-40fc-9a81-5f002c9421bc {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 720.921964] env[61987]: DEBUG nova.compute.manager [req-6f125fcc-e7c9-42de-997a-570902b4250e req-e5cbb9b0-a0a0-4946-b37e-a403db2d0a19 service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Received event network-vif-plugged-b3fafafe-77c6-41b9-ae56-b17b2d781bca {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 720.922329] env[61987]: DEBUG oslo_concurrency.lockutils [req-6f125fcc-e7c9-42de-997a-570902b4250e req-e5cbb9b0-a0a0-4946-b37e-a403db2d0a19 service nova] Acquiring lock "93174a10-f8b1-4789-ab3c-dda07bdddaa8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.922745] env[61987]: DEBUG oslo_concurrency.lockutils [req-6f125fcc-e7c9-42de-997a-570902b4250e req-e5cbb9b0-a0a0-4946-b37e-a403db2d0a19 service nova] Lock "93174a10-f8b1-4789-ab3c-dda07bdddaa8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.923115] env[61987]: DEBUG oslo_concurrency.lockutils [req-6f125fcc-e7c9-42de-997a-570902b4250e req-e5cbb9b0-a0a0-4946-b37e-a403db2d0a19 service nova] Lock "93174a10-f8b1-4789-ab3c-dda07bdddaa8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.923513] env[61987]: DEBUG nova.compute.manager [req-6f125fcc-e7c9-42de-997a-570902b4250e req-e5cbb9b0-a0a0-4946-b37e-a403db2d0a19 service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] No waiting events found dispatching network-vif-plugged-b3fafafe-77c6-41b9-ae56-b17b2d781bca {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 720.923881] env[61987]: WARNING nova.compute.manager [req-6f125fcc-e7c9-42de-997a-570902b4250e req-e5cbb9b0-a0a0-4946-b37e-a403db2d0a19 service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Received unexpected event network-vif-plugged-b3fafafe-77c6-41b9-ae56-b17b2d781bca for instance with vm_state building and task_state spawning. [ 720.924233] env[61987]: DEBUG nova.compute.manager [req-6f125fcc-e7c9-42de-997a-570902b4250e req-e5cbb9b0-a0a0-4946-b37e-a403db2d0a19 service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Received event network-changed-b3fafafe-77c6-41b9-ae56-b17b2d781bca {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 720.924598] env[61987]: DEBUG nova.compute.manager [req-6f125fcc-e7c9-42de-997a-570902b4250e req-e5cbb9b0-a0a0-4946-b37e-a403db2d0a19 service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Refreshing instance network info cache due to event network-changed-b3fafafe-77c6-41b9-ae56-b17b2d781bca. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 720.925118] env[61987]: DEBUG oslo_concurrency.lockutils [req-6f125fcc-e7c9-42de-997a-570902b4250e req-e5cbb9b0-a0a0-4946-b37e-a403db2d0a19 service nova] Acquiring lock "refresh_cache-93174a10-f8b1-4789-ab3c-dda07bdddaa8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.925414] env[61987]: DEBUG oslo_concurrency.lockutils [req-6f125fcc-e7c9-42de-997a-570902b4250e req-e5cbb9b0-a0a0-4946-b37e-a403db2d0a19 service nova] Acquired lock "refresh_cache-93174a10-f8b1-4789-ab3c-dda07bdddaa8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.925688] env[61987]: DEBUG nova.network.neutron [req-6f125fcc-e7c9-42de-997a-570902b4250e req-e5cbb9b0-a0a0-4946-b37e-a403db2d0a19 service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Refreshing network info cache for port b3fafafe-77c6-41b9-ae56-b17b2d781bca {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 721.120473] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.527s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.120473] env[61987]: DEBUG nova.compute.manager [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 721.121732] env[61987]: DEBUG oslo_concurrency.lockutils [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.871s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.123318] env[61987]: INFO nova.compute.claims [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 721.137876] env[61987]: INFO nova.compute.manager [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] instance snapshotting [ 721.139265] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f4a950-d46f-45dd-b2c6-2d00186978b9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.161548] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bbdd862-de61-446b-8eac-b181acfe7074 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.240272] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061402, 'name': Rename_Task, 'duration_secs': 0.166753} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.240272] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 721.240272] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7b53fe64-9a13-4d79-81a1-aeffb5df6e0c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.248267] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for the task: (returnval){ [ 721.248267] env[61987]: value = "task-1061405" [ 721.248267] env[61987]: _type = "Task" [ 721.248267] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.257409] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061405, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.329033] env[61987]: DEBUG oslo_vmware.api [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061404, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151458} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.329033] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 721.329033] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 721.329033] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 721.329033] env[61987]: INFO nova.compute.manager [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Took 1.12 seconds to destroy the instance on the hypervisor. [ 721.329734] env[61987]: DEBUG oslo.service.loopingcall [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 721.329734] env[61987]: DEBUG nova.compute.manager [-] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 721.329734] env[61987]: DEBUG nova.network.neutron [-] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 721.343020] env[61987]: INFO nova.compute.manager [-] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Took 1.68 seconds to deallocate network for instance. [ 721.354625] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': task-1061400, 'name': Rename_Task, 'duration_secs': 1.119306} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.355242] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 721.355639] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a915181c-95c2-42cc-8deb-202264940b94 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.362170] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Waiting for the task: (returnval){ [ 721.362170] env[61987]: value = "task-1061406" [ 721.362170] env[61987]: _type = "Task" [ 721.362170] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.371094] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': task-1061406, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.612846] env[61987]: DEBUG nova.network.neutron [req-6f125fcc-e7c9-42de-997a-570902b4250e req-e5cbb9b0-a0a0-4946-b37e-a403db2d0a19 service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 721.635563] env[61987]: DEBUG nova.compute.utils [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 721.643245] env[61987]: DEBUG nova.compute.manager [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 721.643517] env[61987]: DEBUG nova.network.neutron [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 721.675897] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Creating Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 721.676354] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8f3b14ef-9748-43e2-993e-36f25aa5be15 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.696519] env[61987]: DEBUG oslo_vmware.api [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for the task: (returnval){ [ 721.696519] env[61987]: value = "task-1061407" [ 721.696519] env[61987]: _type = "Task" [ 721.696519] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.709250] env[61987]: DEBUG oslo_vmware.api [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061407, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.728841] env[61987]: DEBUG nova.policy [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7d15929a29cb498485c2b3bf77da9457', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d8a522ba9bf422db98ca69c8d92b373', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 721.761626] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061405, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.850915] env[61987]: DEBUG oslo_concurrency.lockutils [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.883417] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': task-1061406, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.002704] env[61987]: DEBUG nova.network.neutron [req-6f125fcc-e7c9-42de-997a-570902b4250e req-e5cbb9b0-a0a0-4946-b37e-a403db2d0a19 service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.118146] env[61987]: DEBUG nova.network.neutron [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Successfully created port: 39309545-c724-44c7-932e-7784c9e31fa7 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 722.147075] env[61987]: DEBUG nova.compute.manager [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 722.207829] env[61987]: DEBUG oslo_vmware.api [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061407, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.258141] env[61987]: DEBUG oslo_vmware.api [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061405, 'name': PowerOnVM_Task, 'duration_secs': 0.578403} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.258438] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 722.258686] env[61987]: INFO nova.compute.manager [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Took 8.91 seconds to spawn the instance on the hypervisor. [ 722.258875] env[61987]: DEBUG nova.compute.manager [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 722.260201] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f25ee367-22cd-406b-abc3-ac11b5755e91 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.376906] env[61987]: DEBUG oslo_vmware.api [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': task-1061406, 'name': PowerOnVM_Task, 'duration_secs': 0.965446} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.377372] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 722.377804] env[61987]: INFO nova.compute.manager [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Took 11.55 seconds to spawn the instance on the hypervisor. [ 722.378139] env[61987]: DEBUG nova.compute.manager [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 722.380489] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e79631c-48b3-4a48-b6cd-65af7e03dcb7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.510821] env[61987]: DEBUG oslo_concurrency.lockutils [req-6f125fcc-e7c9-42de-997a-570902b4250e req-e5cbb9b0-a0a0-4946-b37e-a403db2d0a19 service nova] Releasing lock "refresh_cache-93174a10-f8b1-4789-ab3c-dda07bdddaa8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.631478] env[61987]: DEBUG nova.network.neutron [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Successfully updated port: 27b82c05-3eb6-48ab-9455-b321ef1f1fe6 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 722.651155] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-078ec244-c851-4617-8e8c-399b5570c412 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.668613] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b9b5d6b-a0be-4cd3-97db-428c1e21a9d5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.674952] env[61987]: DEBUG nova.network.neutron [-] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.705454] env[61987]: INFO nova.compute.manager [-] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Took 1.38 seconds to deallocate network for instance. [ 722.709232] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd59440-bf97-4614-8c5c-63ae3ec0c995 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.723218] env[61987]: DEBUG oslo_vmware.api [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061407, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.724799] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed0e0007-5900-4fe5-ad72-54d2be0523c2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.741455] env[61987]: DEBUG nova.compute.provider_tree [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.785895] env[61987]: INFO nova.compute.manager [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Took 26.20 seconds to build instance. [ 722.896570] env[61987]: INFO nova.compute.manager [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Took 28.69 seconds to build instance. [ 722.978799] env[61987]: DEBUG nova.network.neutron [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Successfully updated port: a4965cbd-fb11-4afb-acec-bf0128f32e0f {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 723.134827] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Acquiring lock "refresh_cache-d55395cc-0b92-4e99-9b80-913c19f6c7f4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.134997] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Acquired lock "refresh_cache-d55395cc-0b92-4e99-9b80-913c19f6c7f4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.135174] env[61987]: DEBUG nova.network.neutron [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 723.166218] env[61987]: DEBUG nova.compute.manager [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 723.201389] env[61987]: DEBUG nova.virt.hardware [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 723.201717] env[61987]: DEBUG nova.virt.hardware [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 723.201888] env[61987]: DEBUG nova.virt.hardware [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 723.202106] env[61987]: DEBUG nova.virt.hardware [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 723.202277] env[61987]: DEBUG nova.virt.hardware [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 723.202425] env[61987]: DEBUG nova.virt.hardware [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 723.202641] env[61987]: DEBUG nova.virt.hardware [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 723.202808] env[61987]: DEBUG nova.virt.hardware [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 723.202982] env[61987]: DEBUG nova.virt.hardware [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 723.203172] env[61987]: DEBUG nova.virt.hardware [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 723.203360] env[61987]: DEBUG nova.virt.hardware [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 723.204305] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a37c018-ce7f-40a8-a736-3fd47bcde8b0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.218734] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ccdca04-67fc-4bfd-98b4-989100b5e2d3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.223867] env[61987]: DEBUG oslo_vmware.api [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061407, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.224756] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.245257] env[61987]: DEBUG nova.scheduler.client.report [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 723.288815] env[61987]: DEBUG oslo_concurrency.lockutils [None req-206500f8-a7bb-43be-9c0b-469861f022b7 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.860s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.399446] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3892774-7a31-4f80-a018-9a77bb157e22 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Lock "e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.041s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.435506] env[61987]: DEBUG nova.compute.manager [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Received event network-vif-deleted-af8b91f3-b342-4408-b3da-5489a317c8b0 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 723.435740] env[61987]: DEBUG nova.compute.manager [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Received event network-vif-plugged-27b82c05-3eb6-48ab-9455-b321ef1f1fe6 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 723.436026] env[61987]: DEBUG oslo_concurrency.lockutils [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] Acquiring lock "d55395cc-0b92-4e99-9b80-913c19f6c7f4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.436290] env[61987]: DEBUG oslo_concurrency.lockutils [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] Lock "d55395cc-0b92-4e99-9b80-913c19f6c7f4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.436515] env[61987]: DEBUG oslo_concurrency.lockutils [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] Lock "d55395cc-0b92-4e99-9b80-913c19f6c7f4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.436647] env[61987]: DEBUG nova.compute.manager [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] No waiting events found dispatching network-vif-plugged-27b82c05-3eb6-48ab-9455-b321ef1f1fe6 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 723.436827] env[61987]: WARNING nova.compute.manager [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Received unexpected event network-vif-plugged-27b82c05-3eb6-48ab-9455-b321ef1f1fe6 for instance with vm_state building and task_state spawning. [ 723.437222] env[61987]: DEBUG nova.compute.manager [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Received event network-vif-deleted-574078f6-ac6b-4540-808d-a4b8a1183854 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 723.437222] env[61987]: DEBUG nova.compute.manager [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Received event network-changed-27b82c05-3eb6-48ab-9455-b321ef1f1fe6 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 723.437345] env[61987]: DEBUG nova.compute.manager [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Refreshing instance network info cache due to event network-changed-27b82c05-3eb6-48ab-9455-b321ef1f1fe6. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 723.437517] env[61987]: DEBUG oslo_concurrency.lockutils [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] Acquiring lock "refresh_cache-d55395cc-0b92-4e99-9b80-913c19f6c7f4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.481591] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Acquiring lock "refresh_cache-93174a10-f8b1-4789-ab3c-dda07bdddaa8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.481735] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Acquired lock "refresh_cache-93174a10-f8b1-4789-ab3c-dda07bdddaa8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.481893] env[61987]: DEBUG nova.network.neutron [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 723.667614] env[61987]: DEBUG nova.network.neutron [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 723.717459] env[61987]: DEBUG oslo_vmware.api [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061407, 'name': CreateSnapshot_Task, 'duration_secs': 1.56218} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.717752] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Created Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 723.718537] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-362402c9-75fc-4c78-bdfd-86a82093564f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.751038] env[61987]: DEBUG oslo_concurrency.lockutils [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.629s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.751575] env[61987]: DEBUG nova.compute.manager [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 723.755354] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.226s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.756015] env[61987]: INFO nova.compute.claims [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 723.791850] env[61987]: DEBUG nova.compute.manager [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 723.893098] env[61987]: DEBUG nova.network.neutron [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Updating instance_info_cache with network_info: [{"id": "27b82c05-3eb6-48ab-9455-b321ef1f1fe6", "address": "fa:16:3e:e7:19:c0", "network": {"id": "f17b381d-4a1c-4a71-b92b-ddd5a5bc2809", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-312450176-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8a22d437ed0e42089681884b0d28c182", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56398cc0-e39f-410f-8036-8c2a6870e26f", "external-id": "nsx-vlan-transportzone-612", "segmentation_id": 612, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27b82c05-3e", "ovs_interfaceid": "27b82c05-3eb6-48ab-9455-b321ef1f1fe6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.903471] env[61987]: DEBUG nova.compute.manager [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 724.025818] env[61987]: DEBUG nova.network.neutron [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.154024] env[61987]: DEBUG nova.compute.manager [req-34c466f5-0c65-4a39-90e3-1685b74f0630 req-2f2738cd-8535-45b5-8b86-c76e9dd5df39 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Received event network-vif-plugged-39309545-c724-44c7-932e-7784c9e31fa7 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 724.154281] env[61987]: DEBUG oslo_concurrency.lockutils [req-34c466f5-0c65-4a39-90e3-1685b74f0630 req-2f2738cd-8535-45b5-8b86-c76e9dd5df39 service nova] Acquiring lock "4edd0159-6db1-41bd-a616-31a34e984059-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.154583] env[61987]: DEBUG oslo_concurrency.lockutils [req-34c466f5-0c65-4a39-90e3-1685b74f0630 req-2f2738cd-8535-45b5-8b86-c76e9dd5df39 service nova] Lock "4edd0159-6db1-41bd-a616-31a34e984059-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.154793] env[61987]: DEBUG oslo_concurrency.lockutils [req-34c466f5-0c65-4a39-90e3-1685b74f0630 req-2f2738cd-8535-45b5-8b86-c76e9dd5df39 service nova] Lock "4edd0159-6db1-41bd-a616-31a34e984059-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.155524] env[61987]: DEBUG nova.compute.manager [req-34c466f5-0c65-4a39-90e3-1685b74f0630 req-2f2738cd-8535-45b5-8b86-c76e9dd5df39 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] No waiting events found dispatching network-vif-plugged-39309545-c724-44c7-932e-7784c9e31fa7 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 724.155823] env[61987]: WARNING nova.compute.manager [req-34c466f5-0c65-4a39-90e3-1685b74f0630 req-2f2738cd-8535-45b5-8b86-c76e9dd5df39 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Received unexpected event network-vif-plugged-39309545-c724-44c7-932e-7784c9e31fa7 for instance with vm_state building and task_state spawning. [ 724.241586] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Creating linked-clone VM from snapshot {{(pid=61987) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 724.244343] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d5dcc68e-b4e0-4e56-961d-61ff028b1582 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.254053] env[61987]: DEBUG oslo_vmware.api [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for the task: (returnval){ [ 724.254053] env[61987]: value = "task-1061408" [ 724.254053] env[61987]: _type = "Task" [ 724.254053] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.264081] env[61987]: DEBUG nova.compute.utils [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 724.267288] env[61987]: DEBUG nova.network.neutron [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Successfully updated port: 39309545-c724-44c7-932e-7784c9e31fa7 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 724.267772] env[61987]: DEBUG nova.compute.manager [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 724.267964] env[61987]: DEBUG nova.network.neutron [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 724.276600] env[61987]: DEBUG oslo_vmware.api [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061408, 'name': CloneVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.320507] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.347020] env[61987]: DEBUG nova.policy [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c124c6b758543a68e141796b585e41f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bd47eb44bc334bf3ae5813905903ecbc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 724.395436] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Releasing lock "refresh_cache-d55395cc-0b92-4e99-9b80-913c19f6c7f4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.395783] env[61987]: DEBUG nova.compute.manager [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Instance network_info: |[{"id": "27b82c05-3eb6-48ab-9455-b321ef1f1fe6", "address": "fa:16:3e:e7:19:c0", "network": {"id": "f17b381d-4a1c-4a71-b92b-ddd5a5bc2809", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-312450176-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8a22d437ed0e42089681884b0d28c182", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56398cc0-e39f-410f-8036-8c2a6870e26f", "external-id": "nsx-vlan-transportzone-612", "segmentation_id": 612, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27b82c05-3e", "ovs_interfaceid": "27b82c05-3eb6-48ab-9455-b321ef1f1fe6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 724.396234] env[61987]: DEBUG oslo_concurrency.lockutils [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] Acquired lock "refresh_cache-d55395cc-0b92-4e99-9b80-913c19f6c7f4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.396500] env[61987]: DEBUG nova.network.neutron [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Refreshing network info cache for port 27b82c05-3eb6-48ab-9455-b321ef1f1fe6 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 724.397991] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e7:19:c0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56398cc0-e39f-410f-8036-8c2a6870e26f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '27b82c05-3eb6-48ab-9455-b321ef1f1fe6', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 724.411046] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Creating folder: Project (8a22d437ed0e42089681884b0d28c182). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 724.414299] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-440dd09e-cfc4-4ac4-a841-c0a66f7385a4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.427589] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Created folder: Project (8a22d437ed0e42089681884b0d28c182) in parent group-v234219. [ 724.427911] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Creating folder: Instances. Parent ref: group-v234277. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 724.428473] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f794d6b1-b0e9-4a78-af5a-858e03ac6dc0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.433953] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.438176] env[61987]: DEBUG nova.network.neutron [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Updating instance_info_cache with network_info: [{"id": "b3fafafe-77c6-41b9-ae56-b17b2d781bca", "address": "fa:16:3e:aa:c7:35", "network": {"id": "edb7f5c0-87ae-4dc5-a3d8-a918a16fa90d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1917160120", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.111", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2287d9f275945a5959acb3e6ef26074", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "664c466b-9417-49d7-83cc-364d964c403a", "external-id": "nsx-vlan-transportzone-103", "segmentation_id": 103, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3fafafe-77", "ovs_interfaceid": "b3fafafe-77c6-41b9-ae56-b17b2d781bca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a4965cbd-fb11-4afb-acec-bf0128f32e0f", "address": "fa:16:3e:99:36:47", "network": {"id": "ea2f2e38-53e8-4c03-a34f-74c41ecc7486", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-328954170", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.251", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "a2287d9f275945a5959acb3e6ef26074", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa4965cbd-fb", "ovs_interfaceid": "a4965cbd-fb11-4afb-acec-bf0128f32e0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.442493] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Created folder: Instances in parent group-v234277. [ 724.443136] env[61987]: DEBUG oslo.service.loopingcall [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 724.443973] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 724.443973] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1d0fc09a-c994-4d30-a08d-c2267d2abd40 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.470655] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 724.470655] env[61987]: value = "task-1061411" [ 724.470655] env[61987]: _type = "Task" [ 724.470655] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.480559] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061411, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.764057] env[61987]: DEBUG nova.network.neutron [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Successfully created port: 853aa7a2-96d2-4665-b680-811dffc26bfe {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 724.774626] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquiring lock "refresh_cache-4edd0159-6db1-41bd-a616-31a34e984059" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.774626] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquired lock "refresh_cache-4edd0159-6db1-41bd-a616-31a34e984059" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.774626] env[61987]: DEBUG nova.network.neutron [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 724.779489] env[61987]: DEBUG nova.compute.manager [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 724.780898] env[61987]: DEBUG oslo_vmware.api [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061408, 'name': CloneVM_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.940645] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Releasing lock "refresh_cache-93174a10-f8b1-4789-ab3c-dda07bdddaa8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.942022] env[61987]: DEBUG nova.compute.manager [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Instance network_info: |[{"id": "b3fafafe-77c6-41b9-ae56-b17b2d781bca", "address": "fa:16:3e:aa:c7:35", "network": {"id": "edb7f5c0-87ae-4dc5-a3d8-a918a16fa90d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1917160120", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.111", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2287d9f275945a5959acb3e6ef26074", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "664c466b-9417-49d7-83cc-364d964c403a", "external-id": "nsx-vlan-transportzone-103", "segmentation_id": 103, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3fafafe-77", "ovs_interfaceid": "b3fafafe-77c6-41b9-ae56-b17b2d781bca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a4965cbd-fb11-4afb-acec-bf0128f32e0f", "address": "fa:16:3e:99:36:47", "network": {"id": "ea2f2e38-53e8-4c03-a34f-74c41ecc7486", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-328954170", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.251", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "a2287d9f275945a5959acb3e6ef26074", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa4965cbd-fb", "ovs_interfaceid": "a4965cbd-fb11-4afb-acec-bf0128f32e0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 724.942641] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:aa:c7:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '664c466b-9417-49d7-83cc-364d964c403a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b3fafafe-77c6-41b9-ae56-b17b2d781bca', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:36:47', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b71230ae-e879-4384-88ce-fe64c86fce22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a4965cbd-fb11-4afb-acec-bf0128f32e0f', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 724.953811] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Creating folder: Project (a2287d9f275945a5959acb3e6ef26074). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 724.962647] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-365d3f85-dc0f-4543-8457-a59e099a147c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.975810] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Created folder: Project (a2287d9f275945a5959acb3e6ef26074) in parent group-v234219. [ 724.975937] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Creating folder: Instances. Parent ref: group-v234280. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 724.976625] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e3e9526f-6e60-4d17-bb14-3b82606940a8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.986256] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061411, 'name': CreateVM_Task, 'duration_secs': 0.377687} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.992033] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 724.992226] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Created folder: Instances in parent group-v234280. [ 724.992981] env[61987]: DEBUG oslo.service.loopingcall [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 724.999706] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.999706] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.999706] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 724.999706] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 724.999706] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b80017a-9fe8-463e-b908-5c247f26a1cf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.001278] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d4216577-bd6f-4e7a-895f-2b3e8467b408 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.026606] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Waiting for the task: (returnval){ [ 725.026606] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5253560a-6632-a82e-85b4-e516d0fe283a" [ 725.026606] env[61987]: _type = "Task" [ 725.026606] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.028690] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 725.028690] env[61987]: value = "task-1061414" [ 725.028690] env[61987]: _type = "Task" [ 725.028690] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.048885] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5253560a-6632-a82e-85b4-e516d0fe283a, 'name': SearchDatastore_Task, 'duration_secs': 0.013111} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.052824] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.052824] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 725.053071] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.053139] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.053317] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 725.053981] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061414, 'name': CreateVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.053981] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1585fed3-f2ce-4c34-85a0-7c0d65efb279 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.078454] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 725.079411] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 725.079606] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-749c41cd-7ba6-4849-8f3a-b11f1d026b39 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.085947] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Waiting for the task: (returnval){ [ 725.085947] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]522c5705-3314-47ab-9023-ee3cfa7ea418" [ 725.085947] env[61987]: _type = "Task" [ 725.085947] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.099501] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]522c5705-3314-47ab-9023-ee3cfa7ea418, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.267549] env[61987]: DEBUG oslo_vmware.api [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061408, 'name': CloneVM_Task} progress is 95%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.310302] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0f179c5-45c0-4cef-aecc-7c7b78f2fae0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.318235] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdfbbd60-6dba-4065-b32b-da03d344b0e5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.353486] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5067a9c6-f655-4676-9ce0-da17bb4d09a0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.362947] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b836ca1a-e6ae-4fde-9373-3fc8ea786a35 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.368381] env[61987]: DEBUG nova.network.neutron [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 725.384865] env[61987]: DEBUG nova.compute.provider_tree [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.407222] env[61987]: DEBUG nova.network.neutron [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Updated VIF entry in instance network info cache for port 27b82c05-3eb6-48ab-9455-b321ef1f1fe6. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 725.407599] env[61987]: DEBUG nova.network.neutron [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Updating instance_info_cache with network_info: [{"id": "27b82c05-3eb6-48ab-9455-b321ef1f1fe6", "address": "fa:16:3e:e7:19:c0", "network": {"id": "f17b381d-4a1c-4a71-b92b-ddd5a5bc2809", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-312450176-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8a22d437ed0e42089681884b0d28c182", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56398cc0-e39f-410f-8036-8c2a6870e26f", "external-id": "nsx-vlan-transportzone-612", "segmentation_id": 612, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27b82c05-3e", "ovs_interfaceid": "27b82c05-3eb6-48ab-9455-b321ef1f1fe6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.464391] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Acquiring lock "e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.464761] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Lock "e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.465404] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Acquiring lock "e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.465615] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Lock "e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.465819] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Lock "e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.474286] env[61987]: INFO nova.compute.manager [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Terminating instance [ 725.545815] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061414, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.600218] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]522c5705-3314-47ab-9023-ee3cfa7ea418, 'name': SearchDatastore_Task, 'duration_secs': 0.017806} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.603459] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3de36b7-f685-477e-b475-fda85d33b3d4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.607647] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Waiting for the task: (returnval){ [ 725.607647] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5287a5bb-36e2-7102-4bc9-199dc4739c36" [ 725.607647] env[61987]: _type = "Task" [ 725.607647] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.618948] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5287a5bb-36e2-7102-4bc9-199dc4739c36, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.626383] env[61987]: DEBUG nova.network.neutron [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Updating instance_info_cache with network_info: [{"id": "39309545-c724-44c7-932e-7784c9e31fa7", "address": "fa:16:3e:fd:80:60", "network": {"id": "90535a3b-74ac-4bdc-a5ea-78272b01cab6", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1880668508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d8a522ba9bf422db98ca69c8d92b373", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39309545-c7", "ovs_interfaceid": "39309545-c724-44c7-932e-7784c9e31fa7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.768499] env[61987]: DEBUG oslo_vmware.api [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061408, 'name': CloneVM_Task, 'duration_secs': 1.376548} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.768890] env[61987]: INFO nova.virt.vmwareapi.vmops [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Created linked-clone VM from snapshot [ 725.769991] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3796391c-c249-4c22-9e47-b0770da93764 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.779357] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Uploading image 0098a037-893f-46c6-816c-12e3250d8b8d {{(pid=61987) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 725.793638] env[61987]: DEBUG nova.compute.manager [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 725.809846] env[61987]: DEBUG nova.compute.manager [req-c187b6b2-dbd7-45e5-baf1-0ef43c58d264 req-c1737792-b4be-4640-8a88-99cc9ab401d3 service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Received event network-changed-a4965cbd-fb11-4afb-acec-bf0128f32e0f {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 725.810023] env[61987]: DEBUG nova.compute.manager [req-c187b6b2-dbd7-45e5-baf1-0ef43c58d264 req-c1737792-b4be-4640-8a88-99cc9ab401d3 service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Refreshing instance network info cache due to event network-changed-a4965cbd-fb11-4afb-acec-bf0128f32e0f. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 725.810199] env[61987]: DEBUG oslo_concurrency.lockutils [req-c187b6b2-dbd7-45e5-baf1-0ef43c58d264 req-c1737792-b4be-4640-8a88-99cc9ab401d3 service nova] Acquiring lock "refresh_cache-93174a10-f8b1-4789-ab3c-dda07bdddaa8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.810389] env[61987]: DEBUG oslo_concurrency.lockutils [req-c187b6b2-dbd7-45e5-baf1-0ef43c58d264 req-c1737792-b4be-4640-8a88-99cc9ab401d3 service nova] Acquired lock "refresh_cache-93174a10-f8b1-4789-ab3c-dda07bdddaa8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.810574] env[61987]: DEBUG nova.network.neutron [req-c187b6b2-dbd7-45e5-baf1-0ef43c58d264 req-c1737792-b4be-4640-8a88-99cc9ab401d3 service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Refreshing network info cache for port a4965cbd-fb11-4afb-acec-bf0128f32e0f {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 725.815826] env[61987]: DEBUG oslo_vmware.rw_handles [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 725.815826] env[61987]: value = "vm-234276" [ 725.815826] env[61987]: _type = "VirtualMachine" [ 725.815826] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 725.817639] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-01663aaa-b2a6-413e-acdf-918f0a695baa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.825666] env[61987]: DEBUG nova.virt.hardware [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 725.825914] env[61987]: DEBUG nova.virt.hardware [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 725.826093] env[61987]: DEBUG nova.virt.hardware [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 725.826292] env[61987]: DEBUG nova.virt.hardware [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 725.826446] env[61987]: DEBUG nova.virt.hardware [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 725.826630] env[61987]: DEBUG nova.virt.hardware [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 725.826890] env[61987]: DEBUG nova.virt.hardware [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 725.827088] env[61987]: DEBUG nova.virt.hardware [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 725.827269] env[61987]: DEBUG nova.virt.hardware [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 725.827436] env[61987]: DEBUG nova.virt.hardware [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 725.827614] env[61987]: DEBUG nova.virt.hardware [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 725.828776] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdeef0d1-988c-4480-99d5-e566b40cda4e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.833144] env[61987]: DEBUG oslo_vmware.rw_handles [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Lease: (returnval){ [ 725.833144] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52d71a8a-5366-7d92-91ff-657aab8b48d1" [ 725.833144] env[61987]: _type = "HttpNfcLease" [ 725.833144] env[61987]: } obtained for exporting VM: (result){ [ 725.833144] env[61987]: value = "vm-234276" [ 725.833144] env[61987]: _type = "VirtualMachine" [ 725.833144] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 725.833435] env[61987]: DEBUG oslo_vmware.api [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for the lease: (returnval){ [ 725.833435] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52d71a8a-5366-7d92-91ff-657aab8b48d1" [ 725.833435] env[61987]: _type = "HttpNfcLease" [ 725.833435] env[61987]: } to be ready. {{(pid=61987) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 725.840687] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af10bd24-3a6f-42c2-afbe-8a996253dd71 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.846832] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 725.846832] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52d71a8a-5366-7d92-91ff-657aab8b48d1" [ 725.846832] env[61987]: _type = "HttpNfcLease" [ 725.846832] env[61987]: } is initializing. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 725.892146] env[61987]: DEBUG nova.scheduler.client.report [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 725.910564] env[61987]: DEBUG oslo_concurrency.lockutils [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] Releasing lock "refresh_cache-d55395cc-0b92-4e99-9b80-913c19f6c7f4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.911629] env[61987]: DEBUG nova.compute.manager [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Received event network-changed-fd844030-513e-4774-a178-8a6203365b9c {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 725.911629] env[61987]: DEBUG nova.compute.manager [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Refreshing instance network info cache due to event network-changed-fd844030-513e-4774-a178-8a6203365b9c. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 725.911629] env[61987]: DEBUG oslo_concurrency.lockutils [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] Acquiring lock "refresh_cache-8ff4ebb9-57ea-49d2-8d3c-5876da61e77b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.911629] env[61987]: DEBUG oslo_concurrency.lockutils [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] Acquired lock "refresh_cache-8ff4ebb9-57ea-49d2-8d3c-5876da61e77b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.911629] env[61987]: DEBUG nova.network.neutron [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Refreshing network info cache for port fd844030-513e-4774-a178-8a6203365b9c {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 725.981213] env[61987]: DEBUG nova.compute.manager [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 725.982027] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 725.982416] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d024d40b-8f45-4538-bd74-0277758956a4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.990776] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 725.991073] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e185c2c-9fa9-41d0-8326-d9fe41ea351a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.998079] env[61987]: DEBUG oslo_vmware.api [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Waiting for the task: (returnval){ [ 725.998079] env[61987]: value = "task-1061416" [ 725.998079] env[61987]: _type = "Task" [ 725.998079] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.009904] env[61987]: DEBUG oslo_vmware.api [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': task-1061416, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.043902] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061414, 'name': CreateVM_Task, 'duration_secs': 0.551861} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.044205] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 726.045118] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.045319] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.045645] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 726.045904] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-089f08c9-6468-4ba2-a985-942f3491c0c8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.050846] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Waiting for the task: (returnval){ [ 726.050846] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52d263b4-36a0-46a1-cb31-e051d93045df" [ 726.050846] env[61987]: _type = "Task" [ 726.050846] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.060417] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52d263b4-36a0-46a1-cb31-e051d93045df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.119083] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5287a5bb-36e2-7102-4bc9-199dc4739c36, 'name': SearchDatastore_Task, 'duration_secs': 0.02261} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.119200] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.119406] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] d55395cc-0b92-4e99-9b80-913c19f6c7f4/d55395cc-0b92-4e99-9b80-913c19f6c7f4.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 726.119692] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b7d5197b-a5cd-466e-8f48-d5b83a6c8d6a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.126944] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Waiting for the task: (returnval){ [ 726.126944] env[61987]: value = "task-1061417" [ 726.126944] env[61987]: _type = "Task" [ 726.126944] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.130459] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Releasing lock "refresh_cache-4edd0159-6db1-41bd-a616-31a34e984059" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.130817] env[61987]: DEBUG nova.compute.manager [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Instance network_info: |[{"id": "39309545-c724-44c7-932e-7784c9e31fa7", "address": "fa:16:3e:fd:80:60", "network": {"id": "90535a3b-74ac-4bdc-a5ea-78272b01cab6", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1880668508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d8a522ba9bf422db98ca69c8d92b373", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39309545-c7", "ovs_interfaceid": "39309545-c724-44c7-932e-7784c9e31fa7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 726.131278] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fd:80:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60567ee6-01d0-4b16-9c7a-4a896827d6eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '39309545-c724-44c7-932e-7784c9e31fa7', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 726.139372] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Creating folder: Project (8d8a522ba9bf422db98ca69c8d92b373). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 726.140164] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0f3b0809-3805-40fd-be7e-21fcea2197cc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.145224] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': task-1061417, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.153414] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Created folder: Project (8d8a522ba9bf422db98ca69c8d92b373) in parent group-v234219. [ 726.153699] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Creating folder: Instances. Parent ref: group-v234283. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 726.153970] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3d4bd031-1474-4fe3-88cf-b50f4f7c5fdd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.162666] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Created folder: Instances in parent group-v234283. [ 726.162980] env[61987]: DEBUG oslo.service.loopingcall [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 726.163205] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 726.163421] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fbfd2626-7819-47ac-8eb2-c3354c8c24dc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.182741] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 726.182741] env[61987]: value = "task-1061420" [ 726.182741] env[61987]: _type = "Task" [ 726.182741] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.190804] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061420, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.343321] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 726.343321] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52d71a8a-5366-7d92-91ff-657aab8b48d1" [ 726.343321] env[61987]: _type = "HttpNfcLease" [ 726.343321] env[61987]: } is ready. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 726.343828] env[61987]: DEBUG oslo_vmware.rw_handles [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 726.343828] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52d71a8a-5366-7d92-91ff-657aab8b48d1" [ 726.343828] env[61987]: _type = "HttpNfcLease" [ 726.343828] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 726.346023] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a59b6ed9-0d03-4d65-8eff-3af73d2e0e64 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.349947] env[61987]: DEBUG nova.compute.manager [req-a79d3c42-163b-4429-9942-ed23d35ca88a req-716d6f98-a943-4ccb-90b7-b58d96f89489 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Received event network-changed-39309545-c724-44c7-932e-7784c9e31fa7 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 726.350163] env[61987]: DEBUG nova.compute.manager [req-a79d3c42-163b-4429-9942-ed23d35ca88a req-716d6f98-a943-4ccb-90b7-b58d96f89489 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Refreshing instance network info cache due to event network-changed-39309545-c724-44c7-932e-7784c9e31fa7. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 726.350553] env[61987]: DEBUG oslo_concurrency.lockutils [req-a79d3c42-163b-4429-9942-ed23d35ca88a req-716d6f98-a943-4ccb-90b7-b58d96f89489 service nova] Acquiring lock "refresh_cache-4edd0159-6db1-41bd-a616-31a34e984059" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.350789] env[61987]: DEBUG oslo_concurrency.lockutils [req-a79d3c42-163b-4429-9942-ed23d35ca88a req-716d6f98-a943-4ccb-90b7-b58d96f89489 service nova] Acquired lock "refresh_cache-4edd0159-6db1-41bd-a616-31a34e984059" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.350994] env[61987]: DEBUG nova.network.neutron [req-a79d3c42-163b-4429-9942-ed23d35ca88a req-716d6f98-a943-4ccb-90b7-b58d96f89489 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Refreshing network info cache for port 39309545-c724-44c7-932e-7784c9e31fa7 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 726.362820] env[61987]: DEBUG oslo_vmware.rw_handles [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52899531-02de-3c63-adb0-f79a7ec521a7/disk-0.vmdk from lease info. {{(pid=61987) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 726.363071] env[61987]: DEBUG oslo_vmware.rw_handles [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52899531-02de-3c63-adb0-f79a7ec521a7/disk-0.vmdk for reading. {{(pid=61987) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 726.433393] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.679s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.434140] env[61987]: DEBUG nova.compute.manager [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 726.444258] env[61987]: DEBUG oslo_concurrency.lockutils [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.382s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.446302] env[61987]: INFO nova.compute.claims [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 726.457608] env[61987]: DEBUG nova.compute.manager [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Stashing vm_state: active {{(pid=61987) _prep_resize /opt/stack/nova/nova/compute/manager.py:6090}} [ 726.511059] env[61987]: DEBUG oslo_vmware.api [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': task-1061416, 'name': PowerOffVM_Task, 'duration_secs': 0.33544} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.511419] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 726.511621] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 726.511929] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a4c66a60-e661-4ae5-a018-b49716ef7b06 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.566286] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52d263b4-36a0-46a1-cb31-e051d93045df, 'name': SearchDatastore_Task, 'duration_secs': 0.015169} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.566943] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.566943] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 726.567311] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.567393] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.567633] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 726.568025] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97eed76d-049b-429c-9f88-18fe92370301 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.583158] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 726.583446] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 726.584472] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ecb867bf-5060-4387-a2ae-b9e9030efc88 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.591613] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Waiting for the task: (returnval){ [ 726.591613] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52616059-7644-d63d-d6ef-5aa15200cec8" [ 726.591613] env[61987]: _type = "Task" [ 726.591613] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.600625] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 726.600898] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 726.601126] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Deleting the datastore file [datastore2] e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 726.602088] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c7a5a892-f7c5-4f9f-b747-3a41f0a41020 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.614401] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52616059-7644-d63d-d6ef-5aa15200cec8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.620220] env[61987]: DEBUG oslo_vmware.api [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Waiting for the task: (returnval){ [ 726.620220] env[61987]: value = "task-1061422" [ 726.620220] env[61987]: _type = "Task" [ 726.620220] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.632126] env[61987]: DEBUG oslo_vmware.api [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': task-1061422, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.643989] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': task-1061417, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.661123] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-efc3af56-7ffc-456a-b68f-c8b8b3dbb80a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.700825] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061420, 'name': CreateVM_Task, 'duration_secs': 0.468932} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.709027] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 726.709027] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.709027] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.709027] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 726.709027] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b92b933-0355-4ed9-913c-0f038b17faa1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.718584] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for the task: (returnval){ [ 726.718584] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52c706c3-4f9b-389b-95ea-d6246e5b581b" [ 726.718584] env[61987]: _type = "Task" [ 726.718584] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.728799] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c706c3-4f9b-389b-95ea-d6246e5b581b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.807697] env[61987]: DEBUG nova.network.neutron [req-c187b6b2-dbd7-45e5-baf1-0ef43c58d264 req-c1737792-b4be-4640-8a88-99cc9ab401d3 service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Updated VIF entry in instance network info cache for port a4965cbd-fb11-4afb-acec-bf0128f32e0f. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 726.807697] env[61987]: DEBUG nova.network.neutron [req-c187b6b2-dbd7-45e5-baf1-0ef43c58d264 req-c1737792-b4be-4640-8a88-99cc9ab401d3 service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Updating instance_info_cache with network_info: [{"id": "b3fafafe-77c6-41b9-ae56-b17b2d781bca", "address": "fa:16:3e:aa:c7:35", "network": {"id": "edb7f5c0-87ae-4dc5-a3d8-a918a16fa90d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1917160120", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.111", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2287d9f275945a5959acb3e6ef26074", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "664c466b-9417-49d7-83cc-364d964c403a", "external-id": "nsx-vlan-transportzone-103", "segmentation_id": 103, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3fafafe-77", "ovs_interfaceid": "b3fafafe-77c6-41b9-ae56-b17b2d781bca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a4965cbd-fb11-4afb-acec-bf0128f32e0f", "address": "fa:16:3e:99:36:47", "network": {"id": "ea2f2e38-53e8-4c03-a34f-74c41ecc7486", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-328954170", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.251", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "a2287d9f275945a5959acb3e6ef26074", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b71230ae-e879-4384-88ce-fe64c86fce22", "external-id": "nsx-vlan-transportzone-473", "segmentation_id": 473, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa4965cbd-fb", "ovs_interfaceid": "a4965cbd-fb11-4afb-acec-bf0128f32e0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.834900] env[61987]: DEBUG nova.network.neutron [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Successfully updated port: 853aa7a2-96d2-4665-b680-811dffc26bfe {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 726.926020] env[61987]: DEBUG nova.network.neutron [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Updated VIF entry in instance network info cache for port fd844030-513e-4774-a178-8a6203365b9c. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 726.926020] env[61987]: DEBUG nova.network.neutron [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Updating instance_info_cache with network_info: [{"id": "fd844030-513e-4774-a178-8a6203365b9c", "address": "fa:16:3e:a8:bd:b9", "network": {"id": "662d2418-580e-4675-974e-8fca3f56aaa1", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1448080498-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.130", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ea973b343034355916bca2cce20a22f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd844030-51", "ovs_interfaceid": "fd844030-513e-4774-a178-8a6203365b9c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.939863] env[61987]: DEBUG nova.compute.utils [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 726.941474] env[61987]: DEBUG nova.compute.manager [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 726.941646] env[61987]: DEBUG nova.network.neutron [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 726.982267] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.990673] env[61987]: DEBUG oslo_vmware.rw_handles [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5253a660-d193-da86-08f5-d70d89a16a13/disk-0.vmdk. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 726.991667] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37af4cb6-f0ba-460f-a2d9-04336ca50eee {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.000999] env[61987]: DEBUG oslo_vmware.rw_handles [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5253a660-d193-da86-08f5-d70d89a16a13/disk-0.vmdk is in state: ready. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 727.001320] env[61987]: ERROR oslo_vmware.rw_handles [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5253a660-d193-da86-08f5-d70d89a16a13/disk-0.vmdk due to incomplete transfer. [ 727.001565] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-79cd935d-f092-4bae-a803-2270d821dd52 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.008253] env[61987]: DEBUG oslo_vmware.rw_handles [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5253a660-d193-da86-08f5-d70d89a16a13/disk-0.vmdk. {{(pid=61987) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 727.008460] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Uploaded image f31f91a0-19e6-447b-abcc-cd75f8cffd89 to the Glance image server {{(pid=61987) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 727.010895] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Destroying the VM {{(pid=61987) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 727.011235] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-7a5b2aa6-90a4-4883-8edc-22d47366e731 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.014165] env[61987]: DEBUG nova.policy [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '750d9bd6f45c4fc3bd2b23c254fba92f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd32dc833164a428dad56d1eeb9a7c3eb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 727.020173] env[61987]: DEBUG oslo_vmware.api [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 727.020173] env[61987]: value = "task-1061423" [ 727.020173] env[61987]: _type = "Task" [ 727.020173] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.028865] env[61987]: DEBUG oslo_vmware.api [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061423, 'name': Destroy_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.101376] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52616059-7644-d63d-d6ef-5aa15200cec8, 'name': SearchDatastore_Task, 'duration_secs': 0.039595} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.102350] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da5b2d55-ce1a-496c-9563-f71eede5660d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.108100] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Waiting for the task: (returnval){ [ 727.108100] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52b46bdd-b180-2460-b7c1-41d9aa6d8be4" [ 727.108100] env[61987]: _type = "Task" [ 727.108100] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.117353] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b46bdd-b180-2460-b7c1-41d9aa6d8be4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.128996] env[61987]: DEBUG oslo_vmware.api [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Task: {'id': task-1061422, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.27361} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.128996] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 727.129474] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 727.129676] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 727.129898] env[61987]: INFO nova.compute.manager [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Took 1.15 seconds to destroy the instance on the hypervisor. [ 727.130215] env[61987]: DEBUG oslo.service.loopingcall [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 727.130420] env[61987]: DEBUG nova.compute.manager [-] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 727.130531] env[61987]: DEBUG nova.network.neutron [-] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 727.142993] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': task-1061417, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.648318} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.143342] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] d55395cc-0b92-4e99-9b80-913c19f6c7f4/d55395cc-0b92-4e99-9b80-913c19f6c7f4.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 727.143659] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 727.144019] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-eb682dd7-b435-4a10-8cac-02118d294987 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.150844] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Waiting for the task: (returnval){ [ 727.150844] env[61987]: value = "task-1061424" [ 727.150844] env[61987]: _type = "Task" [ 727.150844] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.161663] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': task-1061424, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.230124] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c706c3-4f9b-389b-95ea-d6246e5b581b, 'name': SearchDatastore_Task, 'duration_secs': 0.044292} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.233612] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.233994] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 727.234342] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.312102] env[61987]: DEBUG nova.network.neutron [req-a79d3c42-163b-4429-9942-ed23d35ca88a req-716d6f98-a943-4ccb-90b7-b58d96f89489 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Updated VIF entry in instance network info cache for port 39309545-c724-44c7-932e-7784c9e31fa7. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 727.312546] env[61987]: DEBUG nova.network.neutron [req-a79d3c42-163b-4429-9942-ed23d35ca88a req-716d6f98-a943-4ccb-90b7-b58d96f89489 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Updating instance_info_cache with network_info: [{"id": "39309545-c724-44c7-932e-7784c9e31fa7", "address": "fa:16:3e:fd:80:60", "network": {"id": "90535a3b-74ac-4bdc-a5ea-78272b01cab6", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1880668508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d8a522ba9bf422db98ca69c8d92b373", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39309545-c7", "ovs_interfaceid": "39309545-c724-44c7-932e-7784c9e31fa7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.317084] env[61987]: DEBUG oslo_concurrency.lockutils [req-c187b6b2-dbd7-45e5-baf1-0ef43c58d264 req-c1737792-b4be-4640-8a88-99cc9ab401d3 service nova] Releasing lock "refresh_cache-93174a10-f8b1-4789-ab3c-dda07bdddaa8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.337314] env[61987]: DEBUG oslo_concurrency.lockutils [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "refresh_cache-325183b9-e35b-4f9c-a271-d0fdb05288bb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.337485] env[61987]: DEBUG oslo_concurrency.lockutils [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired lock "refresh_cache-325183b9-e35b-4f9c-a271-d0fdb05288bb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.337751] env[61987]: DEBUG nova.network.neutron [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 727.428710] env[61987]: DEBUG oslo_concurrency.lockutils [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] Releasing lock "refresh_cache-8ff4ebb9-57ea-49d2-8d3c-5876da61e77b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.429046] env[61987]: DEBUG nova.compute.manager [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Received event network-vif-plugged-a4965cbd-fb11-4afb-acec-bf0128f32e0f {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 727.430791] env[61987]: DEBUG oslo_concurrency.lockutils [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] Acquiring lock "93174a10-f8b1-4789-ab3c-dda07bdddaa8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.430791] env[61987]: DEBUG oslo_concurrency.lockutils [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] Lock "93174a10-f8b1-4789-ab3c-dda07bdddaa8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.430791] env[61987]: DEBUG oslo_concurrency.lockutils [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] Lock "93174a10-f8b1-4789-ab3c-dda07bdddaa8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.430791] env[61987]: DEBUG nova.compute.manager [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] No waiting events found dispatching network-vif-plugged-a4965cbd-fb11-4afb-acec-bf0128f32e0f {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 727.430791] env[61987]: WARNING nova.compute.manager [req-4531ef8c-854c-4002-8581-7344315007cd req-f80f04b9-d11a-4202-9fc6-4645e8991226 service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Received unexpected event network-vif-plugged-a4965cbd-fb11-4afb-acec-bf0128f32e0f for instance with vm_state building and task_state spawning. [ 727.448382] env[61987]: DEBUG nova.compute.manager [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 727.534120] env[61987]: DEBUG oslo_vmware.api [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061423, 'name': Destroy_Task} progress is 33%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.632281] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b46bdd-b180-2460-b7c1-41d9aa6d8be4, 'name': SearchDatastore_Task, 'duration_secs': 0.016276} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.632924] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.633740] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 93174a10-f8b1-4789-ab3c-dda07bdddaa8/93174a10-f8b1-4789-ab3c-dda07bdddaa8.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 727.634548] env[61987]: DEBUG nova.network.neutron [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Successfully created port: 203d9e60-9c1f-4c51-9ee8-9a34be19d289 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 727.636532] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.638334] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 727.638334] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-167eaffd-aece-4589-b5d2-621a3387b7ec {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.641647] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-12a4cba2-bba2-46d2-b8a6-f1fdc03d14c7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.650173] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Waiting for the task: (returnval){ [ 727.650173] env[61987]: value = "task-1061425" [ 727.650173] env[61987]: _type = "Task" [ 727.650173] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.657304] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 727.657577] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 727.663160] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-adeac0df-19d1-4625-9420-11f399b752e0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.669035] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': task-1061425, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.675301] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': task-1061424, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.246691} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.679481] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 727.680303] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for the task: (returnval){ [ 727.680303] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52b5ecef-7300-25e3-54cf-13c8a23d5d5f" [ 727.680303] env[61987]: _type = "Task" [ 727.680303] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.681255] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c17413-9da6-4ae3-a97a-3ea5a341d8a6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.713278] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] d55395cc-0b92-4e99-9b80-913c19f6c7f4/d55395cc-0b92-4e99-9b80-913c19f6c7f4.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 727.713278] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b5ecef-7300-25e3-54cf-13c8a23d5d5f, 'name': SearchDatastore_Task, 'duration_secs': 0.01377} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.716663] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-771dfa0b-6d70-4644-88cf-a5925e1965fe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.733197] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b18714f-7d30-4214-bbcf-5e03ec8e9e05 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.741315] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for the task: (returnval){ [ 727.741315] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5264cf40-62b7-b72a-2a9d-d2cf16777f2b" [ 727.741315] env[61987]: _type = "Task" [ 727.741315] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.745825] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Waiting for the task: (returnval){ [ 727.745825] env[61987]: value = "task-1061426" [ 727.745825] env[61987]: _type = "Task" [ 727.745825] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.753319] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5264cf40-62b7-b72a-2a9d-d2cf16777f2b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.758215] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': task-1061426, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.816034] env[61987]: DEBUG oslo_concurrency.lockutils [req-a79d3c42-163b-4429-9942-ed23d35ca88a req-716d6f98-a943-4ccb-90b7-b58d96f89489 service nova] Releasing lock "refresh_cache-4edd0159-6db1-41bd-a616-31a34e984059" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.895750] env[61987]: DEBUG nova.network.neutron [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 728.022344] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0cc99b5-48c3-48ff-9201-c6ed24bea3fa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.036831] env[61987]: DEBUG oslo_vmware.api [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061423, 'name': Destroy_Task, 'duration_secs': 0.919805} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.039315] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Destroyed the VM [ 728.039700] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Deleting Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 728.040118] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-8a4ae564-9d06-41af-8703-7a39181f2162 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.042846] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-741ac00f-551c-4ca2-b87b-06266af36dac {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.078974] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b9428d3-d53a-45cf-a4f0-4a78ab34f19d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.082083] env[61987]: DEBUG oslo_vmware.api [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 728.082083] env[61987]: value = "task-1061427" [ 728.082083] env[61987]: _type = "Task" [ 728.082083] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.095056] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a6e6c3-7f80-4a96-b150-9688b20005b6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.102930] env[61987]: DEBUG oslo_vmware.api [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061427, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.115324] env[61987]: DEBUG nova.compute.provider_tree [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 728.142500] env[61987]: DEBUG nova.network.neutron [-] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.165529] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': task-1061425, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.233047] env[61987]: DEBUG nova.network.neutron [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Updating instance_info_cache with network_info: [{"id": "853aa7a2-96d2-4665-b680-811dffc26bfe", "address": "fa:16:3e:62:4a:9c", "network": {"id": "ca68bf96-3f18-45ee-80fb-64a1260ed996", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-733595377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd47eb44bc334bf3ae5813905903ecbc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a27fd90b-16a5-43af-bede-ae36762ece00", "external-id": "nsx-vlan-transportzone-197", "segmentation_id": 197, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap853aa7a2-96", "ovs_interfaceid": "853aa7a2-96d2-4665-b680-811dffc26bfe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.254259] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5264cf40-62b7-b72a-2a9d-d2cf16777f2b, 'name': SearchDatastore_Task, 'duration_secs': 0.017868} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.255042] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.257026] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 4edd0159-6db1-41bd-a616-31a34e984059/4edd0159-6db1-41bd-a616-31a34e984059.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 728.257026] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9691dab0-d4cd-4281-9e00-7505a77cd433 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.262444] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': task-1061426, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.268115] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for the task: (returnval){ [ 728.268115] env[61987]: value = "task-1061428" [ 728.268115] env[61987]: _type = "Task" [ 728.268115] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.283657] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061428, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.413316] env[61987]: DEBUG nova.compute.manager [req-74b58695-c720-4612-b94c-4ffaeccb08be req-2fb5d3e5-7d16-4da2-9415-32932af6642a service nova] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Received event network-vif-deleted-f0fc4290-34ae-42ea-9704-e52bdcfdee85 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 728.462386] env[61987]: DEBUG nova.compute.manager [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 728.489909] env[61987]: DEBUG nova.virt.hardware [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:09:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1272212075',id=38,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1232703482',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 728.490362] env[61987]: DEBUG nova.virt.hardware [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 728.490612] env[61987]: DEBUG nova.virt.hardware [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 728.490824] env[61987]: DEBUG nova.virt.hardware [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 728.490984] env[61987]: DEBUG nova.virt.hardware [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 728.492170] env[61987]: DEBUG nova.virt.hardware [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 728.492401] env[61987]: DEBUG nova.virt.hardware [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 728.492758] env[61987]: DEBUG nova.virt.hardware [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 728.493135] env[61987]: DEBUG nova.virt.hardware [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 728.493450] env[61987]: DEBUG nova.virt.hardware [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 728.493819] env[61987]: DEBUG nova.virt.hardware [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 728.494835] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89b9d94f-c6ef-40fb-a898-711e39a94ffe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.503551] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1ab6ace-a1be-45a5-ad27-99f91c478d5d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.594625] env[61987]: DEBUG oslo_vmware.api [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061427, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.619179] env[61987]: DEBUG nova.scheduler.client.report [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 728.645156] env[61987]: INFO nova.compute.manager [-] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Took 1.51 seconds to deallocate network for instance. [ 728.662988] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': task-1061425, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.760575} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.663537] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 93174a10-f8b1-4789-ab3c-dda07bdddaa8/93174a10-f8b1-4789-ab3c-dda07bdddaa8.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 728.663815] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 728.664159] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6713d7a7-361e-4d69-9d82-6c8ada91bc35 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.675492] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Waiting for the task: (returnval){ [ 728.675492] env[61987]: value = "task-1061429" [ 728.675492] env[61987]: _type = "Task" [ 728.675492] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.689844] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': task-1061429, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.738044] env[61987]: DEBUG oslo_concurrency.lockutils [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Releasing lock "refresh_cache-325183b9-e35b-4f9c-a271-d0fdb05288bb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.738516] env[61987]: DEBUG nova.compute.manager [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Instance network_info: |[{"id": "853aa7a2-96d2-4665-b680-811dffc26bfe", "address": "fa:16:3e:62:4a:9c", "network": {"id": "ca68bf96-3f18-45ee-80fb-64a1260ed996", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-733595377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd47eb44bc334bf3ae5813905903ecbc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a27fd90b-16a5-43af-bede-ae36762ece00", "external-id": "nsx-vlan-transportzone-197", "segmentation_id": 197, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap853aa7a2-96", "ovs_interfaceid": "853aa7a2-96d2-4665-b680-811dffc26bfe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 728.739077] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:4a:9c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a27fd90b-16a5-43af-bede-ae36762ece00', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '853aa7a2-96d2-4665-b680-811dffc26bfe', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 728.749147] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Creating folder: Project (bd47eb44bc334bf3ae5813905903ecbc). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 728.749864] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-39dc2da2-7a58-494c-ab48-6530026f08a6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.761793] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': task-1061426, 'name': ReconfigVM_Task, 'duration_secs': 0.709122} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.762274] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Reconfigured VM instance instance-0000002d to attach disk [datastore2] d55395cc-0b92-4e99-9b80-913c19f6c7f4/d55395cc-0b92-4e99-9b80-913c19f6c7f4.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 728.763024] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4e9f8f8c-8f73-4788-9aac-8a55e214922b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.766909] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Created folder: Project (bd47eb44bc334bf3ae5813905903ecbc) in parent group-v234219. [ 728.767295] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Creating folder: Instances. Parent ref: group-v234286. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 728.768071] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-223d9b47-3a57-4898-b689-5e445d72be25 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.776108] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Waiting for the task: (returnval){ [ 728.776108] env[61987]: value = "task-1061431" [ 728.776108] env[61987]: _type = "Task" [ 728.776108] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.786318] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061428, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.786625] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Created folder: Instances in parent group-v234286. [ 728.786864] env[61987]: DEBUG oslo.service.loopingcall [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 728.790011] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 728.790361] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': task-1061431, 'name': Rename_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.790669] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0c83900e-a622-43f4-a832-03f9314ba1b9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.816139] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 728.816139] env[61987]: value = "task-1061433" [ 728.816139] env[61987]: _type = "Task" [ 728.816139] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.825213] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061433, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.867131] env[61987]: DEBUG nova.compute.manager [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Received event network-vif-plugged-853aa7a2-96d2-4665-b680-811dffc26bfe {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 728.867543] env[61987]: DEBUG oslo_concurrency.lockutils [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] Acquiring lock "325183b9-e35b-4f9c-a271-d0fdb05288bb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.867674] env[61987]: DEBUG oslo_concurrency.lockutils [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] Lock "325183b9-e35b-4f9c-a271-d0fdb05288bb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.867809] env[61987]: DEBUG oslo_concurrency.lockutils [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] Lock "325183b9-e35b-4f9c-a271-d0fdb05288bb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.868189] env[61987]: DEBUG nova.compute.manager [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] No waiting events found dispatching network-vif-plugged-853aa7a2-96d2-4665-b680-811dffc26bfe {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 728.868373] env[61987]: WARNING nova.compute.manager [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Received unexpected event network-vif-plugged-853aa7a2-96d2-4665-b680-811dffc26bfe for instance with vm_state building and task_state spawning. [ 728.868558] env[61987]: DEBUG nova.compute.manager [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Received event network-changed-853aa7a2-96d2-4665-b680-811dffc26bfe {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 728.868888] env[61987]: DEBUG nova.compute.manager [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Refreshing instance network info cache due to event network-changed-853aa7a2-96d2-4665-b680-811dffc26bfe. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 728.869075] env[61987]: DEBUG oslo_concurrency.lockutils [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] Acquiring lock "refresh_cache-325183b9-e35b-4f9c-a271-d0fdb05288bb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.869349] env[61987]: DEBUG oslo_concurrency.lockutils [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] Acquired lock "refresh_cache-325183b9-e35b-4f9c-a271-d0fdb05288bb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.869521] env[61987]: DEBUG nova.network.neutron [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Refreshing network info cache for port 853aa7a2-96d2-4665-b680-811dffc26bfe {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 729.094807] env[61987]: DEBUG oslo_vmware.api [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061427, 'name': RemoveSnapshot_Task, 'duration_secs': 0.698669} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.095167] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Deleted Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 729.095414] env[61987]: INFO nova.compute.manager [None req-1f06e90b-ebb8-4e31-baf9-9c7d31408525 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Took 16.69 seconds to snapshot the instance on the hypervisor. [ 729.126850] env[61987]: DEBUG oslo_concurrency.lockutils [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.682s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.127148] env[61987]: DEBUG nova.compute.manager [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 729.131292] env[61987]: DEBUG oslo_concurrency.lockutils [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.356s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.132968] env[61987]: INFO nova.compute.claims [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 729.151235] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.183284] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': task-1061429, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.103413} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.183735] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 729.184701] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01254e1f-ef90-478a-8765-3156ba815430 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.211281] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] 93174a10-f8b1-4789-ab3c-dda07bdddaa8/93174a10-f8b1-4789-ab3c-dda07bdddaa8.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 729.211637] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-11dff708-e742-4668-80de-8f99fa83c1cc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.232420] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Waiting for the task: (returnval){ [ 729.232420] env[61987]: value = "task-1061434" [ 729.232420] env[61987]: _type = "Task" [ 729.232420] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.240928] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': task-1061434, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.280085] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061428, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.715163} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.283302] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 4edd0159-6db1-41bd-a616-31a34e984059/4edd0159-6db1-41bd-a616-31a34e984059.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 729.283535] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 729.283840] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f646717e-a319-4e39-859d-5e1503f1aab3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.292712] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': task-1061431, 'name': Rename_Task, 'duration_secs': 0.251681} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.294224] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 729.294596] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for the task: (returnval){ [ 729.294596] env[61987]: value = "task-1061435" [ 729.294596] env[61987]: _type = "Task" [ 729.294596] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.294814] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e0da651-f32e-4449-bd93-3903031872b5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.305607] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061435, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.307050] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Waiting for the task: (returnval){ [ 729.307050] env[61987]: value = "task-1061436" [ 729.307050] env[61987]: _type = "Task" [ 729.307050] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.314815] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': task-1061436, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.326856] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061433, 'name': CreateVM_Task, 'duration_secs': 0.357615} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.327093] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 729.327805] env[61987]: DEBUG oslo_concurrency.lockutils [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.328040] env[61987]: DEBUG oslo_concurrency.lockutils [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.328405] env[61987]: DEBUG oslo_concurrency.lockutils [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 729.328702] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b83eb18a-d6d0-4fc5-a15b-667be22fa1c6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.333281] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 729.333281] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]520382fa-5519-3b56-6b6b-17ca3a517060" [ 729.333281] env[61987]: _type = "Task" [ 729.333281] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.344037] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]520382fa-5519-3b56-6b6b-17ca3a517060, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.639022] env[61987]: DEBUG nova.compute.utils [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 729.641455] env[61987]: DEBUG nova.compute.manager [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 729.641817] env[61987]: DEBUG nova.network.neutron [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 729.721184] env[61987]: DEBUG nova.policy [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '45ceca37c2624c6aab775e5052a9597a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c7efd6c745f547d3803d5a6184547bda', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 729.745678] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': task-1061434, 'name': ReconfigVM_Task, 'duration_secs': 0.408954} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.747030] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Reconfigured VM instance instance-0000002c to attach disk [datastore2] 93174a10-f8b1-4789-ab3c-dda07bdddaa8/93174a10-f8b1-4789-ab3c-dda07bdddaa8.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 729.747030] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a4df4245-2294-4ff4-9a42-63dd7bea7f00 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.757023] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Waiting for the task: (returnval){ [ 729.757023] env[61987]: value = "task-1061437" [ 729.757023] env[61987]: _type = "Task" [ 729.757023] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.765304] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': task-1061437, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.811068] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061435, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088162} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.818207] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 729.818207] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6b1f941-1b4d-405b-a94b-5b092fa7c69f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.825541] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': task-1061436, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.844345] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] 4edd0159-6db1-41bd-a616-31a34e984059/4edd0159-6db1-41bd-a616-31a34e984059.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 729.848107] env[61987]: DEBUG oslo_concurrency.lockutils [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "02fe490b-75c8-406b-98e0-aa200f249673" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.848403] env[61987]: DEBUG oslo_concurrency.lockutils [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "02fe490b-75c8-406b-98e0-aa200f249673" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.848625] env[61987]: DEBUG oslo_concurrency.lockutils [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "02fe490b-75c8-406b-98e0-aa200f249673-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.852021] env[61987]: DEBUG oslo_concurrency.lockutils [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "02fe490b-75c8-406b-98e0-aa200f249673-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.852021] env[61987]: DEBUG oslo_concurrency.lockutils [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "02fe490b-75c8-406b-98e0-aa200f249673-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.852021] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f633211-324e-45b0-893a-148ccdcddfd8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.866630] env[61987]: DEBUG nova.network.neutron [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Updated VIF entry in instance network info cache for port 853aa7a2-96d2-4665-b680-811dffc26bfe. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 729.867051] env[61987]: DEBUG nova.network.neutron [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Updating instance_info_cache with network_info: [{"id": "853aa7a2-96d2-4665-b680-811dffc26bfe", "address": "fa:16:3e:62:4a:9c", "network": {"id": "ca68bf96-3f18-45ee-80fb-64a1260ed996", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-733595377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd47eb44bc334bf3ae5813905903ecbc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a27fd90b-16a5-43af-bede-ae36762ece00", "external-id": "nsx-vlan-transportzone-197", "segmentation_id": 197, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap853aa7a2-96", "ovs_interfaceid": "853aa7a2-96d2-4665-b680-811dffc26bfe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.868517] env[61987]: INFO nova.compute.manager [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Terminating instance [ 729.876837] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]520382fa-5519-3b56-6b6b-17ca3a517060, 'name': SearchDatastore_Task, 'duration_secs': 0.010506} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.882343] env[61987]: DEBUG oslo_concurrency.lockutils [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.882343] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 729.882343] env[61987]: DEBUG oslo_concurrency.lockutils [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.882343] env[61987]: DEBUG oslo_concurrency.lockutils [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.882343] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 729.882343] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for the task: (returnval){ [ 729.882343] env[61987]: value = "task-1061438" [ 729.882343] env[61987]: _type = "Task" [ 729.882343] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.882343] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-56082523-de96-47df-a441-f348f6c394f2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.893873] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061438, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.895369] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 729.895541] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 729.896804] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51d63c12-6f3d-4cb5-86ed-f0afda399e1c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.905501] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 729.905501] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]522ba067-0039-415b-6412-ded651e4da87" [ 729.905501] env[61987]: _type = "Task" [ 729.905501] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.914536] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]522ba067-0039-415b-6412-ded651e4da87, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.008645] env[61987]: DEBUG nova.network.neutron [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Successfully updated port: 203d9e60-9c1f-4c51-9ee8-9a34be19d289 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 730.146508] env[61987]: DEBUG nova.compute.manager [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 730.180157] env[61987]: DEBUG nova.network.neutron [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Successfully created port: 6aa03b99-d58f-461e-b2a8-c4552c1d418a {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 730.266896] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': task-1061437, 'name': Rename_Task, 'duration_secs': 0.192866} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.266896] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 730.267162] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fddf4c69-7254-4bb8-b000-37fe7fae4cee {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.275696] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Waiting for the task: (returnval){ [ 730.275696] env[61987]: value = "task-1061439" [ 730.275696] env[61987]: _type = "Task" [ 730.275696] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.287582] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': task-1061439, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.327263] env[61987]: DEBUG oslo_vmware.api [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': task-1061436, 'name': PowerOnVM_Task, 'duration_secs': 0.597388} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.329478] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 730.329772] env[61987]: INFO nova.compute.manager [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Took 9.71 seconds to spawn the instance on the hypervisor. [ 730.330016] env[61987]: DEBUG nova.compute.manager [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 730.331392] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85bee61c-a401-46b0-a624-76d5170e61bd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.377033] env[61987]: DEBUG oslo_concurrency.lockutils [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] Releasing lock "refresh_cache-325183b9-e35b-4f9c-a271-d0fdb05288bb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.377301] env[61987]: DEBUG nova.compute.manager [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Received event network-changed-fd844030-513e-4774-a178-8a6203365b9c {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 730.377767] env[61987]: DEBUG nova.compute.manager [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Refreshing instance network info cache due to event network-changed-fd844030-513e-4774-a178-8a6203365b9c. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 730.378297] env[61987]: DEBUG oslo_concurrency.lockutils [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] Acquiring lock "refresh_cache-8ff4ebb9-57ea-49d2-8d3c-5876da61e77b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.378475] env[61987]: DEBUG oslo_concurrency.lockutils [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] Acquired lock "refresh_cache-8ff4ebb9-57ea-49d2-8d3c-5876da61e77b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.378653] env[61987]: DEBUG nova.network.neutron [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Refreshing network info cache for port fd844030-513e-4774-a178-8a6203365b9c {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 730.383420] env[61987]: DEBUG nova.compute.manager [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 730.383771] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 730.385507] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f7d196b-0ad9-46e3-80f8-581ead2c196a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.401926] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 730.405678] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-557fe146-cbd2-41ab-85f2-3bfd005f35a4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.410573] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061438, 'name': ReconfigVM_Task, 'duration_secs': 0.483419} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.412344] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Reconfigured VM instance instance-0000002e to attach disk [datastore2] 4edd0159-6db1-41bd-a616-31a34e984059/4edd0159-6db1-41bd-a616-31a34e984059.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 730.413990] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9c851d14-f007-47bb-93b7-c48599396c6e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.423174] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]522ba067-0039-415b-6412-ded651e4da87, 'name': SearchDatastore_Task, 'duration_secs': 0.01297} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.428301] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for the task: (returnval){ [ 730.428301] env[61987]: value = "task-1061441" [ 730.428301] env[61987]: _type = "Task" [ 730.428301] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.428635] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ebbcf0a5-a52e-4e65-a186-c90f08cea596 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.443884] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 730.443884] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]520776e5-84fd-6158-8191-868001597d78" [ 730.443884] env[61987]: _type = "Task" [ 730.443884] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.444214] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061441, 'name': Rename_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.460127] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]520776e5-84fd-6158-8191-868001597d78, 'name': SearchDatastore_Task, 'duration_secs': 0.012114} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.463752] env[61987]: DEBUG oslo_concurrency.lockutils [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.464089] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 325183b9-e35b-4f9c-a271-d0fdb05288bb/325183b9-e35b-4f9c-a271-d0fdb05288bb.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 730.464747] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d6fc3f0a-f055-48b4-91fd-c7020e94ce45 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.473176] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 730.473176] env[61987]: value = "task-1061442" [ 730.473176] env[61987]: _type = "Task" [ 730.473176] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.483771] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 730.484295] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 730.484715] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Deleting the datastore file [datastore1] 02fe490b-75c8-406b-98e0-aa200f249673 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 730.486145] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9fa7b54e-a5ea-468a-9802-d874bfbdc9b6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.491925] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061442, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.501018] env[61987]: DEBUG oslo_vmware.api [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 730.501018] env[61987]: value = "task-1061443" [ 730.501018] env[61987]: _type = "Task" [ 730.501018] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.513763] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquiring lock "refresh_cache-d038391c-2b32-455c-85d5-68221e7bd065" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.513948] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquired lock "refresh_cache-d038391c-2b32-455c-85d5-68221e7bd065" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.514157] env[61987]: DEBUG nova.network.neutron [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 730.520292] env[61987]: DEBUG oslo_vmware.api [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061443, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.636544] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6c9c265-9284-496a-990c-34a105bc7014 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.644423] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f25a186e-f171-467c-86d0-7ffad734e83c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.680729] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ecfbc5f-d676-46b7-bd52-12c1d8f1bfe2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.688971] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6af713d-adcb-4c0e-b02a-31f5e6852408 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.704903] env[61987]: DEBUG nova.compute.provider_tree [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.775536] env[61987]: DEBUG nova.compute.manager [req-c184e49d-265e-4196-bad1-63b938a29214 req-c5d6229a-bade-45d1-9fb2-da915a539512 service nova] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Received event network-vif-plugged-203d9e60-9c1f-4c51-9ee8-9a34be19d289 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 730.775536] env[61987]: DEBUG oslo_concurrency.lockutils [req-c184e49d-265e-4196-bad1-63b938a29214 req-c5d6229a-bade-45d1-9fb2-da915a539512 service nova] Acquiring lock "d038391c-2b32-455c-85d5-68221e7bd065-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.775536] env[61987]: DEBUG oslo_concurrency.lockutils [req-c184e49d-265e-4196-bad1-63b938a29214 req-c5d6229a-bade-45d1-9fb2-da915a539512 service nova] Lock "d038391c-2b32-455c-85d5-68221e7bd065-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.775536] env[61987]: DEBUG oslo_concurrency.lockutils [req-c184e49d-265e-4196-bad1-63b938a29214 req-c5d6229a-bade-45d1-9fb2-da915a539512 service nova] Lock "d038391c-2b32-455c-85d5-68221e7bd065-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.775536] env[61987]: DEBUG nova.compute.manager [req-c184e49d-265e-4196-bad1-63b938a29214 req-c5d6229a-bade-45d1-9fb2-da915a539512 service nova] [instance: d038391c-2b32-455c-85d5-68221e7bd065] No waiting events found dispatching network-vif-plugged-203d9e60-9c1f-4c51-9ee8-9a34be19d289 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 730.775536] env[61987]: WARNING nova.compute.manager [req-c184e49d-265e-4196-bad1-63b938a29214 req-c5d6229a-bade-45d1-9fb2-da915a539512 service nova] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Received unexpected event network-vif-plugged-203d9e60-9c1f-4c51-9ee8-9a34be19d289 for instance with vm_state building and task_state spawning. [ 730.775536] env[61987]: DEBUG nova.compute.manager [req-c184e49d-265e-4196-bad1-63b938a29214 req-c5d6229a-bade-45d1-9fb2-da915a539512 service nova] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Received event network-changed-203d9e60-9c1f-4c51-9ee8-9a34be19d289 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 730.775536] env[61987]: DEBUG nova.compute.manager [req-c184e49d-265e-4196-bad1-63b938a29214 req-c5d6229a-bade-45d1-9fb2-da915a539512 service nova] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Refreshing instance network info cache due to event network-changed-203d9e60-9c1f-4c51-9ee8-9a34be19d289. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 730.775825] env[61987]: DEBUG oslo_concurrency.lockutils [req-c184e49d-265e-4196-bad1-63b938a29214 req-c5d6229a-bade-45d1-9fb2-da915a539512 service nova] Acquiring lock "refresh_cache-d038391c-2b32-455c-85d5-68221e7bd065" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.786307] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': task-1061439, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.814514] env[61987]: DEBUG oslo_concurrency.lockutils [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "30c4fff0-c77e-4c11-aced-a040003a6b10" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.814771] env[61987]: DEBUG oslo_concurrency.lockutils [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "30c4fff0-c77e-4c11-aced-a040003a6b10" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.862122] env[61987]: INFO nova.compute.manager [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Took 23.48 seconds to build instance. [ 730.949548] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061441, 'name': Rename_Task, 'duration_secs': 0.199423} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.950100] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 730.950635] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cb5c5036-af37-4928-81e7-bb0a504b8779 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.958489] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for the task: (returnval){ [ 730.958489] env[61987]: value = "task-1061444" [ 730.958489] env[61987]: _type = "Task" [ 730.958489] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.968278] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061444, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.984805] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061442, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.013138] env[61987]: DEBUG oslo_concurrency.lockutils [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Acquiring lock "8ff4ebb9-57ea-49d2-8d3c-5876da61e77b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.013550] env[61987]: DEBUG oslo_concurrency.lockutils [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Lock "8ff4ebb9-57ea-49d2-8d3c-5876da61e77b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.014020] env[61987]: DEBUG oslo_concurrency.lockutils [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Acquiring lock "8ff4ebb9-57ea-49d2-8d3c-5876da61e77b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.014176] env[61987]: DEBUG oslo_concurrency.lockutils [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Lock "8ff4ebb9-57ea-49d2-8d3c-5876da61e77b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.014425] env[61987]: DEBUG oslo_concurrency.lockutils [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Lock "8ff4ebb9-57ea-49d2-8d3c-5876da61e77b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.016892] env[61987]: DEBUG oslo_vmware.api [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061443, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.018935] env[61987]: INFO nova.compute.manager [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Terminating instance [ 731.091304] env[61987]: DEBUG nova.network.neutron [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 731.185397] env[61987]: DEBUG nova.compute.manager [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 731.209068] env[61987]: DEBUG nova.scheduler.client.report [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 731.222229] env[61987]: DEBUG nova.virt.hardware [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 731.222642] env[61987]: DEBUG nova.virt.hardware [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 731.222922] env[61987]: DEBUG nova.virt.hardware [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 731.223920] env[61987]: DEBUG nova.virt.hardware [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 731.224210] env[61987]: DEBUG nova.virt.hardware [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 731.224718] env[61987]: DEBUG nova.virt.hardware [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 731.225105] env[61987]: DEBUG nova.virt.hardware [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 731.225394] env[61987]: DEBUG nova.virt.hardware [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 731.225700] env[61987]: DEBUG nova.virt.hardware [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 731.226009] env[61987]: DEBUG nova.virt.hardware [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 731.226336] env[61987]: DEBUG nova.virt.hardware [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 731.228039] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85968f6a-f88f-4182-aa17-60e5098e44d1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.243944] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f44cdabb-9992-4c0e-9f5f-5aa3a6829c36 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.287573] env[61987]: DEBUG oslo_vmware.api [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': task-1061439, 'name': PowerOnVM_Task, 'duration_secs': 0.616598} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.287944] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 731.288181] env[61987]: INFO nova.compute.manager [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Took 13.34 seconds to spawn the instance on the hypervisor. [ 731.288368] env[61987]: DEBUG nova.compute.manager [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 731.289166] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-280392c9-baf8-4cdc-adf3-c3de25d6501e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.319070] env[61987]: DEBUG nova.network.neutron [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Updated VIF entry in instance network info cache for port fd844030-513e-4774-a178-8a6203365b9c. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 731.319442] env[61987]: DEBUG nova.network.neutron [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Updating instance_info_cache with network_info: [{"id": "fd844030-513e-4774-a178-8a6203365b9c", "address": "fa:16:3e:a8:bd:b9", "network": {"id": "662d2418-580e-4675-974e-8fca3f56aaa1", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-1448080498-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7ea973b343034355916bca2cce20a22f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "49b5df12-d801-4140-8816-2fd401608c7d", "external-id": "nsx-vlan-transportzone-326", "segmentation_id": 326, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd844030-51", "ovs_interfaceid": "fd844030-513e-4774-a178-8a6203365b9c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.362427] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b0f417c2-10ad-4101-a404-b55e20b1bbcc tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Lock "d55395cc-0b92-4e99-9b80-913c19f6c7f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.209s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.391055] env[61987]: DEBUG nova.network.neutron [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Updating instance_info_cache with network_info: [{"id": "203d9e60-9c1f-4c51-9ee8-9a34be19d289", "address": "fa:16:3e:f1:83:58", "network": {"id": "35c53c0f-af9d-4934-84c7-ea4c5252fa58", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-343092073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d32dc833164a428dad56d1eeb9a7c3eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "790c811b-3fa6-49f8-87ac-c51450911137", "external-id": "nsx-vlan-transportzone-908", "segmentation_id": 908, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap203d9e60-9c", "ovs_interfaceid": "203d9e60-9c1f-4c51-9ee8-9a34be19d289", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.469337] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061444, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.485024] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061442, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.771917} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.485024] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 325183b9-e35b-4f9c-a271-d0fdb05288bb/325183b9-e35b-4f9c-a271-d0fdb05288bb.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 731.485319] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 731.485614] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c1b491f3-e2f5-4c3b-b50f-2bffb3e1af04 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.493873] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 731.493873] env[61987]: value = "task-1061445" [ 731.493873] env[61987]: _type = "Task" [ 731.493873] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.504795] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061445, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.513926] env[61987]: DEBUG oslo_vmware.api [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061443, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.531988} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.514320] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 731.514611] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 731.514849] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 731.515061] env[61987]: INFO nova.compute.manager [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Took 1.13 seconds to destroy the instance on the hypervisor. [ 731.515404] env[61987]: DEBUG oslo.service.loopingcall [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 731.515636] env[61987]: DEBUG nova.compute.manager [-] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 731.515737] env[61987]: DEBUG nova.network.neutron [-] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 731.526696] env[61987]: DEBUG nova.compute.manager [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 731.527029] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 731.528828] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20ddccc4-f532-4ed9-ac2d-74b1c5ec7f0b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.537197] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 731.537508] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-32031ef0-d86e-4926-803d-633af44663d0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.544359] env[61987]: DEBUG oslo_vmware.api [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Waiting for the task: (returnval){ [ 731.544359] env[61987]: value = "task-1061446" [ 731.544359] env[61987]: _type = "Task" [ 731.544359] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.552745] env[61987]: DEBUG oslo_vmware.api [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': task-1061446, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.741025] env[61987]: DEBUG oslo_concurrency.lockutils [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.607s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.741025] env[61987]: DEBUG nova.compute.manager [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 731.744093] env[61987]: DEBUG oslo_concurrency.lockutils [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.369s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.744391] env[61987]: DEBUG nova.objects.instance [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Lazy-loading 'resources' on Instance uuid 0b4b3bdb-7393-47ab-abc1-44c47496cb3a {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 731.816883] env[61987]: INFO nova.compute.manager [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Took 30.70 seconds to build instance. [ 731.823026] env[61987]: DEBUG oslo_concurrency.lockutils [req-7c53a9ae-c47e-417f-866a-14ba2227b070 req-f929bf64-a336-4dfe-948d-9606fb932ba0 service nova] Releasing lock "refresh_cache-8ff4ebb9-57ea-49d2-8d3c-5876da61e77b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.868833] env[61987]: DEBUG nova.compute.manager [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 731.879374] env[61987]: DEBUG nova.compute.manager [req-235ab832-4dbf-4024-9dfe-1377cb692594 req-720c4e3e-a2d4-439a-9873-2497493dff76 service nova] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Received event network-vif-plugged-6aa03b99-d58f-461e-b2a8-c4552c1d418a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 731.879374] env[61987]: DEBUG oslo_concurrency.lockutils [req-235ab832-4dbf-4024-9dfe-1377cb692594 req-720c4e3e-a2d4-439a-9873-2497493dff76 service nova] Acquiring lock "ad40937f-d5a9-4708-8b81-06087d38d765-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.879374] env[61987]: DEBUG oslo_concurrency.lockutils [req-235ab832-4dbf-4024-9dfe-1377cb692594 req-720c4e3e-a2d4-439a-9873-2497493dff76 service nova] Lock "ad40937f-d5a9-4708-8b81-06087d38d765-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.879374] env[61987]: DEBUG oslo_concurrency.lockutils [req-235ab832-4dbf-4024-9dfe-1377cb692594 req-720c4e3e-a2d4-439a-9873-2497493dff76 service nova] Lock "ad40937f-d5a9-4708-8b81-06087d38d765-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.879374] env[61987]: DEBUG nova.compute.manager [req-235ab832-4dbf-4024-9dfe-1377cb692594 req-720c4e3e-a2d4-439a-9873-2497493dff76 service nova] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] No waiting events found dispatching network-vif-plugged-6aa03b99-d58f-461e-b2a8-c4552c1d418a {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 731.879374] env[61987]: WARNING nova.compute.manager [req-235ab832-4dbf-4024-9dfe-1377cb692594 req-720c4e3e-a2d4-439a-9873-2497493dff76 service nova] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Received unexpected event network-vif-plugged-6aa03b99-d58f-461e-b2a8-c4552c1d418a for instance with vm_state building and task_state spawning. [ 731.894157] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Releasing lock "refresh_cache-d038391c-2b32-455c-85d5-68221e7bd065" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.894157] env[61987]: DEBUG nova.compute.manager [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Instance network_info: |[{"id": "203d9e60-9c1f-4c51-9ee8-9a34be19d289", "address": "fa:16:3e:f1:83:58", "network": {"id": "35c53c0f-af9d-4934-84c7-ea4c5252fa58", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-343092073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d32dc833164a428dad56d1eeb9a7c3eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "790c811b-3fa6-49f8-87ac-c51450911137", "external-id": "nsx-vlan-transportzone-908", "segmentation_id": 908, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap203d9e60-9c", "ovs_interfaceid": "203d9e60-9c1f-4c51-9ee8-9a34be19d289", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 731.894358] env[61987]: DEBUG oslo_concurrency.lockutils [req-c184e49d-265e-4196-bad1-63b938a29214 req-c5d6229a-bade-45d1-9fb2-da915a539512 service nova] Acquired lock "refresh_cache-d038391c-2b32-455c-85d5-68221e7bd065" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.894596] env[61987]: DEBUG nova.network.neutron [req-c184e49d-265e-4196-bad1-63b938a29214 req-c5d6229a-bade-45d1-9fb2-da915a539512 service nova] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Refreshing network info cache for port 203d9e60-9c1f-4c51-9ee8-9a34be19d289 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 731.899646] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f1:83:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '790c811b-3fa6-49f8-87ac-c51450911137', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '203d9e60-9c1f-4c51-9ee8-9a34be19d289', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 731.911373] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Creating folder: Project (d32dc833164a428dad56d1eeb9a7c3eb). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 731.911734] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e1d62547-f080-4abe-9ddb-28df82424e89 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.923784] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Created folder: Project (d32dc833164a428dad56d1eeb9a7c3eb) in parent group-v234219. [ 731.924044] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Creating folder: Instances. Parent ref: group-v234289. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 731.924294] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cfa4376c-92e0-45a4-b3ab-0392360caa33 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.934325] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Created folder: Instances in parent group-v234289. [ 731.934582] env[61987]: DEBUG oslo.service.loopingcall [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 731.934771] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 731.935019] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-00aab78c-4206-4e2d-9e16-a85a4311b219 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.954802] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 731.954802] env[61987]: value = "task-1061449" [ 731.954802] env[61987]: _type = "Task" [ 731.954802] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.965814] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061449, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.970970] env[61987]: DEBUG oslo_vmware.api [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061444, 'name': PowerOnVM_Task, 'duration_secs': 0.792752} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.971235] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 731.971470] env[61987]: INFO nova.compute.manager [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Took 8.81 seconds to spawn the instance on the hypervisor. [ 731.971680] env[61987]: DEBUG nova.compute.manager [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 731.972465] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d79ebdcc-5ff0-4eb4-b68b-33d76ab4c54f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.975606] env[61987]: DEBUG nova.network.neutron [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Successfully updated port: 6aa03b99-d58f-461e-b2a8-c4552c1d418a {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 732.005224] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061445, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074391} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.005224] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 732.005224] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6d39d11-5549-4a98-ad06-3055f7e99b40 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.030385] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] 325183b9-e35b-4f9c-a271-d0fdb05288bb/325183b9-e35b-4f9c-a271-d0fdb05288bb.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 732.030385] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72914c9a-8c5c-42c8-85e4-638cffd6c1c3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.054992] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 732.054992] env[61987]: value = "task-1061450" [ 732.054992] env[61987]: _type = "Task" [ 732.054992] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.061701] env[61987]: DEBUG oslo_vmware.api [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': task-1061446, 'name': PowerOffVM_Task, 'duration_secs': 0.210928} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.062355] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 732.062535] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 732.062807] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c4a66c29-3d17-4466-aecb-88a459e7ef54 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.067468] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061450, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.120741] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 732.120980] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 732.121193] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Deleting the datastore file [datastore2] 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 732.121470] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-594c36bb-c21e-4c2f-91f3-33627572d451 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.127613] env[61987]: DEBUG oslo_vmware.api [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Waiting for the task: (returnval){ [ 732.127613] env[61987]: value = "task-1061452" [ 732.127613] env[61987]: _type = "Task" [ 732.127613] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.135597] env[61987]: DEBUG oslo_vmware.api [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': task-1061452, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.250526] env[61987]: DEBUG nova.compute.utils [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 732.252130] env[61987]: DEBUG nova.compute.manager [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Not allocating networking since 'none' was specified. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 732.320682] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5727d03b-9f9d-4a65-a3e5-a6af2da18e94 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Lock "93174a10-f8b1-4789-ab3c-dda07bdddaa8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.807s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.355245] env[61987]: DEBUG nova.network.neutron [-] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.392326] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.470358] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061449, 'name': CreateVM_Task, 'duration_secs': 0.333563} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.470694] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 732.474648] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.475626] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.476975] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 732.476975] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7170e0ee-f357-42d1-9054-8af6486403d6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.479052] env[61987]: DEBUG oslo_concurrency.lockutils [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "refresh_cache-ad40937f-d5a9-4708-8b81-06087d38d765" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.479260] env[61987]: DEBUG oslo_concurrency.lockutils [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquired lock "refresh_cache-ad40937f-d5a9-4708-8b81-06087d38d765" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.479442] env[61987]: DEBUG nova.network.neutron [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 732.490942] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 732.490942] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5258254e-5da8-bd98-0462-19af90574a15" [ 732.490942] env[61987]: _type = "Task" [ 732.490942] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.500233] env[61987]: INFO nova.compute.manager [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Took 24.09 seconds to build instance. [ 732.506648] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5258254e-5da8-bd98-0462-19af90574a15, 'name': SearchDatastore_Task, 'duration_secs': 0.013469} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.506987] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.507638] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 732.508798] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.508798] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.508798] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 732.508798] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b3745218-fea7-4779-90c5-93ff018e3fbf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.522840] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 732.522840] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 732.522840] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6575141f-60c3-4bbf-b9e8-9bbd0ca840b8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.531709] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 732.531709] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52f73413-cc03-0fc6-ae2e-510c89e0224c" [ 732.531709] env[61987]: _type = "Task" [ 732.531709] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.546493] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52f73413-cc03-0fc6-ae2e-510c89e0224c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.566206] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061450, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.642981] env[61987]: DEBUG oslo_vmware.api [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Task: {'id': task-1061452, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.26126} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.646468] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 732.646688] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 732.646890] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 732.647101] env[61987]: INFO nova.compute.manager [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Took 1.12 seconds to destroy the instance on the hypervisor. [ 732.647362] env[61987]: DEBUG oslo.service.loopingcall [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 732.648662] env[61987]: DEBUG nova.compute.manager [-] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 732.648662] env[61987]: DEBUG nova.network.neutron [-] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 732.720483] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Acquiring lock "93174a10-f8b1-4789-ab3c-dda07bdddaa8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.720759] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Lock "93174a10-f8b1-4789-ab3c-dda07bdddaa8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.720975] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Acquiring lock "93174a10-f8b1-4789-ab3c-dda07bdddaa8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.721210] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Lock "93174a10-f8b1-4789-ab3c-dda07bdddaa8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.721402] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Lock "93174a10-f8b1-4789-ab3c-dda07bdddaa8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.723708] env[61987]: INFO nova.compute.manager [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Terminating instance [ 732.743222] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92777696-435d-4ec9-850c-9bee7ec4e434 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.745823] env[61987]: DEBUG nova.network.neutron [req-c184e49d-265e-4196-bad1-63b938a29214 req-c5d6229a-bade-45d1-9fb2-da915a539512 service nova] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Updated VIF entry in instance network info cache for port 203d9e60-9c1f-4c51-9ee8-9a34be19d289. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 732.746193] env[61987]: DEBUG nova.network.neutron [req-c184e49d-265e-4196-bad1-63b938a29214 req-c5d6229a-bade-45d1-9fb2-da915a539512 service nova] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Updating instance_info_cache with network_info: [{"id": "203d9e60-9c1f-4c51-9ee8-9a34be19d289", "address": "fa:16:3e:f1:83:58", "network": {"id": "35c53c0f-af9d-4934-84c7-ea4c5252fa58", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-343092073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d32dc833164a428dad56d1eeb9a7c3eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "790c811b-3fa6-49f8-87ac-c51450911137", "external-id": "nsx-vlan-transportzone-908", "segmentation_id": 908, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap203d9e60-9c", "ovs_interfaceid": "203d9e60-9c1f-4c51-9ee8-9a34be19d289", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.752577] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe3c6ae6-dbba-4547-8ebd-89434032ebc7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.756185] env[61987]: DEBUG nova.compute.manager [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 732.789375] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f16b33-7f03-4e86-a16f-d5f45f89558d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.798387] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ece5640-25a4-4500-95d4-5e9cdbefef3b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.815609] env[61987]: DEBUG nova.compute.provider_tree [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.822955] env[61987]: DEBUG nova.compute.manager [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 732.857823] env[61987]: INFO nova.compute.manager [-] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Took 1.34 seconds to deallocate network for instance. [ 733.001325] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4a0d9afa-da47-43bc-bf5f-fca4fa31d438 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "4edd0159-6db1-41bd-a616-31a34e984059" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.570s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.036867] env[61987]: DEBUG nova.network.neutron [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 733.044820] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52f73413-cc03-0fc6-ae2e-510c89e0224c, 'name': SearchDatastore_Task, 'duration_secs': 0.014405} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.046553] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa907b84-5a90-40df-9599-f445e22d0983 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.054268] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 733.054268] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5255df12-230b-37dd-5712-11eac1ada9d7" [ 733.054268] env[61987]: _type = "Task" [ 733.054268] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.068138] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5255df12-230b-37dd-5712-11eac1ada9d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.071906] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061450, 'name': ReconfigVM_Task, 'duration_secs': 0.61537} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.072354] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Reconfigured VM instance instance-0000002f to attach disk [datastore1] 325183b9-e35b-4f9c-a271-d0fdb05288bb/325183b9-e35b-4f9c-a271-d0fdb05288bb.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 733.072848] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-747273d7-d342-44e6-813f-213a5cdc8547 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.078164] env[61987]: DEBUG nova.compute.manager [req-00169fb6-5d8d-411f-ae80-c5211e4a1699 req-3954d1ae-b89a-464f-b5fb-76851b403d6c service nova] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Received event network-vif-deleted-a55b478b-de12-47cb-80e8-07a467463422 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 733.079849] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 733.079849] env[61987]: value = "task-1061453" [ 733.079849] env[61987]: _type = "Task" [ 733.079849] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.089611] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061453, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.228325] env[61987]: DEBUG nova.compute.manager [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 733.228614] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 733.229488] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f0bf0c-3a1b-4d9e-913c-505ae1ec9b53 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.238289] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 733.238567] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-50cd7317-4c84-40f3-9a3d-18bd3c004a41 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.245542] env[61987]: DEBUG oslo_vmware.api [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Waiting for the task: (returnval){ [ 733.245542] env[61987]: value = "task-1061454" [ 733.245542] env[61987]: _type = "Task" [ 733.245542] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.249149] env[61987]: DEBUG oslo_concurrency.lockutils [req-c184e49d-265e-4196-bad1-63b938a29214 req-c5d6229a-bade-45d1-9fb2-da915a539512 service nova] Releasing lock "refresh_cache-d038391c-2b32-455c-85d5-68221e7bd065" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.254523] env[61987]: DEBUG oslo_vmware.api [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': task-1061454, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.295890] env[61987]: DEBUG nova.network.neutron [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Updating instance_info_cache with network_info: [{"id": "6aa03b99-d58f-461e-b2a8-c4552c1d418a", "address": "fa:16:3e:5e:3e:15", "network": {"id": "a9093c43-6bc4-45b1-bdcf-a434759300bd", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-355385599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7efd6c745f547d3803d5a6184547bda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac563aa7-6d7c-4bd5-9241-7b3e11b8f22d", "external-id": "nsx-vlan-transportzone-730", "segmentation_id": 730, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6aa03b99-d5", "ovs_interfaceid": "6aa03b99-d58f-461e-b2a8-c4552c1d418a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.320114] env[61987]: DEBUG nova.scheduler.client.report [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 733.347681] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.365238] env[61987]: DEBUG oslo_concurrency.lockutils [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.478203] env[61987]: DEBUG nova.network.neutron [-] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.505544] env[61987]: DEBUG nova.compute.manager [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 733.566449] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5255df12-230b-37dd-5712-11eac1ada9d7, 'name': SearchDatastore_Task, 'duration_secs': 0.023159} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.566763] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.567106] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] d038391c-2b32-455c-85d5-68221e7bd065/d038391c-2b32-455c-85d5-68221e7bd065.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 733.567687] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7287cef9-78f4-4a99-8103-6ff5af4214ef {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.574861] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 733.574861] env[61987]: value = "task-1061455" [ 733.574861] env[61987]: _type = "Task" [ 733.574861] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.584197] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061455, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.592459] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061453, 'name': Rename_Task, 'duration_secs': 0.233045} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.592757] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 733.593030] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-51f4559e-773b-43d4-8c6d-a7edb71ec1c8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.598385] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 733.598385] env[61987]: value = "task-1061456" [ 733.598385] env[61987]: _type = "Task" [ 733.598385] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.607411] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061456, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.759045] env[61987]: DEBUG oslo_vmware.api [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': task-1061454, 'name': PowerOffVM_Task, 'duration_secs': 0.248436} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.759045] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 733.759509] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 733.759509] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2081e2b5-419f-4b22-84a5-6e93abb67440 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.765693] env[61987]: DEBUG nova.compute.manager [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 733.789161] env[61987]: DEBUG nova.virt.hardware [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 733.789432] env[61987]: DEBUG nova.virt.hardware [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 733.789597] env[61987]: DEBUG nova.virt.hardware [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 733.789791] env[61987]: DEBUG nova.virt.hardware [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 733.789943] env[61987]: DEBUG nova.virt.hardware [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 733.790111] env[61987]: DEBUG nova.virt.hardware [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 733.790335] env[61987]: DEBUG nova.virt.hardware [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 733.790495] env[61987]: DEBUG nova.virt.hardware [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 733.790670] env[61987]: DEBUG nova.virt.hardware [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 733.790833] env[61987]: DEBUG nova.virt.hardware [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 733.791027] env[61987]: DEBUG nova.virt.hardware [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 733.791919] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75581ef0-3646-447d-ac3d-8d8803399dc6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.799485] env[61987]: DEBUG oslo_concurrency.lockutils [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Releasing lock "refresh_cache-ad40937f-d5a9-4708-8b81-06087d38d765" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.799810] env[61987]: DEBUG nova.compute.manager [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Instance network_info: |[{"id": "6aa03b99-d58f-461e-b2a8-c4552c1d418a", "address": "fa:16:3e:5e:3e:15", "network": {"id": "a9093c43-6bc4-45b1-bdcf-a434759300bd", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-355385599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7efd6c745f547d3803d5a6184547bda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac563aa7-6d7c-4bd5-9241-7b3e11b8f22d", "external-id": "nsx-vlan-transportzone-730", "segmentation_id": 730, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6aa03b99-d5", "ovs_interfaceid": "6aa03b99-d58f-461e-b2a8-c4552c1d418a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 733.800296] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:3e:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac563aa7-6d7c-4bd5-9241-7b3e11b8f22d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6aa03b99-d58f-461e-b2a8-c4552c1d418a', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 733.807897] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Creating folder: Project (c7efd6c745f547d3803d5a6184547bda). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 733.809190] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60269bd8-d414-499a-a55f-bbe42d6d9eb9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.813540] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-25dafd5f-279b-4b6c-83ae-865758dfcc34 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.828195] env[61987]: DEBUG oslo_concurrency.lockutils [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.084s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.830517] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Instance VIF info [] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 733.837088] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Creating folder: Project (644af8399cf545adbce85fc5aef33e72). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 733.839152] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.066s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.840851] env[61987]: INFO nova.compute.claims [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 733.843640] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fd454bf5-39a4-41a1-8b9b-64b52019af62 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.845687] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Created folder: Project (c7efd6c745f547d3803d5a6184547bda) in parent group-v234219. [ 733.845788] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Creating folder: Instances. Parent ref: group-v234292. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 733.847042] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1445a7f2-9d67-41f3-be04-e274ebee1941 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.856735] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Created folder: Instances in parent group-v234292. [ 733.858520] env[61987]: DEBUG oslo.service.loopingcall [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 733.858520] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 733.858520] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Created folder: Project (644af8399cf545adbce85fc5aef33e72) in parent group-v234219. [ 733.858689] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Creating folder: Instances. Parent ref: group-v234293. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 733.858847] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ccc8f2ed-bf01-46bd-9eca-546798a4403a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.875231] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1d98bc30-d5fa-4066-8698-b4682ae64744 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.877948] env[61987]: INFO nova.scheduler.client.report [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Deleted allocations for instance 0b4b3bdb-7393-47ab-abc1-44c47496cb3a [ 733.886033] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 733.886033] env[61987]: value = "task-1061462" [ 733.886033] env[61987]: _type = "Task" [ 733.886033] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.891718] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Created folder: Instances in parent group-v234293. [ 733.892021] env[61987]: DEBUG oslo.service.loopingcall [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 733.892255] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 733.892886] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6930051e-5321-4b4d-8ad9-fcf3050eaa0f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.909300] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061462, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.917165] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 733.917165] env[61987]: value = "task-1061463" [ 733.917165] env[61987]: _type = "Task" [ 733.917165] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.927822] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061463, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.929246] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 733.929578] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 733.929691] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Deleting the datastore file [datastore2] 93174a10-f8b1-4789-ab3c-dda07bdddaa8 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 733.929933] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-57a42dd3-d645-4392-9327-e3e589fb8f1b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.938234] env[61987]: DEBUG oslo_vmware.api [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Waiting for the task: (returnval){ [ 733.938234] env[61987]: value = "task-1061464" [ 733.938234] env[61987]: _type = "Task" [ 733.938234] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.951785] env[61987]: DEBUG oslo_vmware.api [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': task-1061464, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.981834] env[61987]: INFO nova.compute.manager [-] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Took 1.33 seconds to deallocate network for instance. [ 734.030757] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.086576] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061455, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.109809] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061456, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.169987] env[61987]: DEBUG nova.compute.manager [req-9abb7de3-7729-41d8-96c6-c005a688e08d req-1ec53e81-447d-4a1f-87ae-3152204c4f58 service nova] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Received event network-changed-6aa03b99-d58f-461e-b2a8-c4552c1d418a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 734.170719] env[61987]: DEBUG nova.compute.manager [req-9abb7de3-7729-41d8-96c6-c005a688e08d req-1ec53e81-447d-4a1f-87ae-3152204c4f58 service nova] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Refreshing instance network info cache due to event network-changed-6aa03b99-d58f-461e-b2a8-c4552c1d418a. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 734.171093] env[61987]: DEBUG oslo_concurrency.lockutils [req-9abb7de3-7729-41d8-96c6-c005a688e08d req-1ec53e81-447d-4a1f-87ae-3152204c4f58 service nova] Acquiring lock "refresh_cache-ad40937f-d5a9-4708-8b81-06087d38d765" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.171894] env[61987]: DEBUG oslo_concurrency.lockutils [req-9abb7de3-7729-41d8-96c6-c005a688e08d req-1ec53e81-447d-4a1f-87ae-3152204c4f58 service nova] Acquired lock "refresh_cache-ad40937f-d5a9-4708-8b81-06087d38d765" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.172166] env[61987]: DEBUG nova.network.neutron [req-9abb7de3-7729-41d8-96c6-c005a688e08d req-1ec53e81-447d-4a1f-87ae-3152204c4f58 service nova] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Refreshing network info cache for port 6aa03b99-d58f-461e-b2a8-c4552c1d418a {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 734.387197] env[61987]: DEBUG oslo_concurrency.lockutils [None req-33069642-7967-407e-84e2-e2a69a8b587c tempest-ServerMetadataTestJSON-2039050955 tempest-ServerMetadataTestJSON-2039050955-project-member] Lock "0b4b3bdb-7393-47ab-abc1-44c47496cb3a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.921s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.398866] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061462, 'name': CreateVM_Task, 'duration_secs': 0.436931} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.400976] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 734.400976] env[61987]: DEBUG oslo_concurrency.lockutils [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.400976] env[61987]: DEBUG oslo_concurrency.lockutils [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.400976] env[61987]: DEBUG oslo_concurrency.lockutils [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 734.400976] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-197619d5-2d4e-41b3-a71f-68fbc1feaf98 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.406659] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 734.406659] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52273a59-5673-ca97-d08f-faf2c41b9a4e" [ 734.406659] env[61987]: _type = "Task" [ 734.406659] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.417743] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52273a59-5673-ca97-d08f-faf2c41b9a4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.426116] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061463, 'name': CreateVM_Task, 'duration_secs': 0.443578} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.426300] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 734.426783] env[61987]: DEBUG oslo_concurrency.lockutils [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.427024] env[61987]: DEBUG oslo_concurrency.lockutils [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.427411] env[61987]: DEBUG oslo_concurrency.lockutils [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 734.427728] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d21effec-5d46-424e-b212-ad52c4ea01ff {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.432698] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for the task: (returnval){ [ 734.432698] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52c1d7d5-fe1e-4969-3c5d-5948493d0b5b" [ 734.432698] env[61987]: _type = "Task" [ 734.432698] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.441450] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c1d7d5-fe1e-4969-3c5d-5948493d0b5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.451095] env[61987]: DEBUG oslo_vmware.api [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Task: {'id': task-1061464, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.435373} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.451095] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 734.451095] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 734.451095] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 734.451274] env[61987]: INFO nova.compute.manager [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Took 1.22 seconds to destroy the instance on the hypervisor. [ 734.452027] env[61987]: DEBUG oslo.service.loopingcall [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 734.452027] env[61987]: DEBUG nova.compute.manager [-] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 734.452027] env[61987]: DEBUG nova.network.neutron [-] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 734.489829] env[61987]: DEBUG oslo_concurrency.lockutils [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.590816] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061455, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.563663} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.591614] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] d038391c-2b32-455c-85d5-68221e7bd065/d038391c-2b32-455c-85d5-68221e7bd065.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 734.591614] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 734.591843] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-689fc7e8-fe74-4865-ac8e-312e7ddc6619 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.600614] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 734.600614] env[61987]: value = "task-1061465" [ 734.600614] env[61987]: _type = "Task" [ 734.600614] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.618920] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061465, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.620523] env[61987]: DEBUG oslo_vmware.api [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061456, 'name': PowerOnVM_Task, 'duration_secs': 0.581999} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.620885] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 734.621067] env[61987]: INFO nova.compute.manager [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Took 8.83 seconds to spawn the instance on the hypervisor. [ 734.621244] env[61987]: DEBUG nova.compute.manager [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 734.622261] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9075de6b-7b45-406b-963f-2e3dc70f5fb4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.787758] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquiring lock "10354ba0-5a64-4174-9c61-152d54a5dfcc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.788102] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "10354ba0-5a64-4174-9c61-152d54a5dfcc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.920421] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52273a59-5673-ca97-d08f-faf2c41b9a4e, 'name': SearchDatastore_Task, 'duration_secs': 0.012603} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.923426] env[61987]: DEBUG oslo_concurrency.lockutils [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.923735] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 734.923991] env[61987]: DEBUG oslo_concurrency.lockutils [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.924164] env[61987]: DEBUG oslo_concurrency.lockutils [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.924371] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 734.925220] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-816880b3-2713-4694-b491-bb2e9510347a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.938044] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 734.938261] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 734.942518] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd8df70d-71a9-4ea6-98ff-66fef0c07c0b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.945040] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c1d7d5-fe1e-4969-3c5d-5948493d0b5b, 'name': SearchDatastore_Task, 'duration_secs': 0.010896} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.947841] env[61987]: DEBUG oslo_concurrency.lockutils [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.948146] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 734.948437] env[61987]: DEBUG oslo_concurrency.lockutils [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.948634] env[61987]: DEBUG oslo_concurrency.lockutils [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.948870] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 734.949736] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6631b7bd-c24b-45df-8103-bb242f8ddd0f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.953230] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 734.953230] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5214b8e6-7ee1-1037-8a2a-ee8020fc22ec" [ 734.953230] env[61987]: _type = "Task" [ 734.953230] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.961920] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5214b8e6-7ee1-1037-8a2a-ee8020fc22ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.970270] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 734.970511] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 734.972730] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-343f70d7-730d-4b22-b9eb-7064bac74662 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.982165] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for the task: (returnval){ [ 734.982165] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5219d8aa-b1ca-08ce-8499-9244f16a691a" [ 734.982165] env[61987]: _type = "Task" [ 734.982165] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.994335] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5219d8aa-b1ca-08ce-8499-9244f16a691a, 'name': SearchDatastore_Task, 'duration_secs': 0.011115} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.999906] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70085279-c5af-4a87-b33a-5b9d370de4a9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.006325] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for the task: (returnval){ [ 735.006325] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52174e72-5df4-a614-0706-7df232747dba" [ 735.006325] env[61987]: _type = "Task" [ 735.006325] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.014256] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52174e72-5df4-a614-0706-7df232747dba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.112847] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061465, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.111858} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.113158] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 735.114215] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c20368f-7ce1-4d6c-a050-3e29a4272b15 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.141829] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] d038391c-2b32-455c-85d5-68221e7bd065/d038391c-2b32-455c-85d5-68221e7bd065.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 735.149056] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af2dbcc3-3db6-45e6-bb5a-85a6c544e0ed {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.167368] env[61987]: INFO nova.compute.manager [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Took 25.95 seconds to build instance. [ 735.172820] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 735.172820] env[61987]: value = "task-1061466" [ 735.172820] env[61987]: _type = "Task" [ 735.172820] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.184727] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061466, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.215995] env[61987]: DEBUG nova.network.neutron [req-9abb7de3-7729-41d8-96c6-c005a688e08d req-1ec53e81-447d-4a1f-87ae-3152204c4f58 service nova] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Updated VIF entry in instance network info cache for port 6aa03b99-d58f-461e-b2a8-c4552c1d418a. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 735.216428] env[61987]: DEBUG nova.network.neutron [req-9abb7de3-7729-41d8-96c6-c005a688e08d req-1ec53e81-447d-4a1f-87ae-3152204c4f58 service nova] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Updating instance_info_cache with network_info: [{"id": "6aa03b99-d58f-461e-b2a8-c4552c1d418a", "address": "fa:16:3e:5e:3e:15", "network": {"id": "a9093c43-6bc4-45b1-bdcf-a434759300bd", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-355385599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7efd6c745f547d3803d5a6184547bda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac563aa7-6d7c-4bd5-9241-7b3e11b8f22d", "external-id": "nsx-vlan-transportzone-730", "segmentation_id": 730, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6aa03b99-d5", "ovs_interfaceid": "6aa03b99-d58f-461e-b2a8-c4552c1d418a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.322750] env[61987]: DEBUG nova.compute.manager [req-1e5e54d5-75b2-4927-9b4c-d6ad98f3a5ff req-4ad9c6e0-3523-42af-aecb-4d2b66ce913b service nova] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Received event network-vif-deleted-fd844030-513e-4774-a178-8a6203365b9c {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 735.322750] env[61987]: DEBUG nova.compute.manager [req-1e5e54d5-75b2-4927-9b4c-d6ad98f3a5ff req-4ad9c6e0-3523-42af-aecb-4d2b66ce913b service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Received event network-vif-deleted-a4965cbd-fb11-4afb-acec-bf0128f32e0f {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 735.322750] env[61987]: INFO nova.compute.manager [req-1e5e54d5-75b2-4927-9b4c-d6ad98f3a5ff req-4ad9c6e0-3523-42af-aecb-4d2b66ce913b service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Neutron deleted interface a4965cbd-fb11-4afb-acec-bf0128f32e0f; detaching it from the instance and deleting it from the info cache [ 735.326321] env[61987]: DEBUG nova.network.neutron [req-1e5e54d5-75b2-4927-9b4c-d6ad98f3a5ff req-4ad9c6e0-3523-42af-aecb-4d2b66ce913b service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Updating instance_info_cache with network_info: [{"id": "b3fafafe-77c6-41b9-ae56-b17b2d781bca", "address": "fa:16:3e:aa:c7:35", "network": {"id": "edb7f5c0-87ae-4dc5-a3d8-a918a16fa90d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1917160120", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.111", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a2287d9f275945a5959acb3e6ef26074", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "664c466b-9417-49d7-83cc-364d964c403a", "external-id": "nsx-vlan-transportzone-103", "segmentation_id": 103, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3fafafe-77", "ovs_interfaceid": "b3fafafe-77c6-41b9-ae56-b17b2d781bca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.350163] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc16846d-2449-4736-bb87-f30b47d85c7e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.359337] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dae1b655-42ec-4f90-90b9-f0e397b213a6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.394565] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c576b58-49f4-4e9a-a707-86cba0ea392b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.406129] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34b0a110-e745-434d-9633-ddaff4846130 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.419411] env[61987]: DEBUG nova.compute.provider_tree [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.471614] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5214b8e6-7ee1-1037-8a2a-ee8020fc22ec, 'name': SearchDatastore_Task, 'duration_secs': 0.012573} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.471614] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6edbde7-734b-4218-9763-9e676f9407a5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.473986] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Acquiring lock "d55395cc-0b92-4e99-9b80-913c19f6c7f4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.474291] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Lock "d55395cc-0b92-4e99-9b80-913c19f6c7f4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.474567] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Acquiring lock "d55395cc-0b92-4e99-9b80-913c19f6c7f4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.474795] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Lock "d55395cc-0b92-4e99-9b80-913c19f6c7f4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.475023] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Lock "d55395cc-0b92-4e99-9b80-913c19f6c7f4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.477509] env[61987]: INFO nova.compute.manager [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Terminating instance [ 735.481628] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 735.481628] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]527d3612-8d77-c427-bcb0-55b2a0bab0fa" [ 735.481628] env[61987]: _type = "Task" [ 735.481628] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.493797] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]527d3612-8d77-c427-bcb0-55b2a0bab0fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.522848] env[61987]: DEBUG nova.network.neutron [-] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.524305] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52174e72-5df4-a614-0706-7df232747dba, 'name': SearchDatastore_Task, 'duration_secs': 0.010422} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.524645] env[61987]: DEBUG oslo_concurrency.lockutils [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.524921] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 1afcffda-0bd2-4cbe-8cfb-12a91bb50975/1afcffda-0bd2-4cbe-8cfb-12a91bb50975.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 735.525467] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a59b25a3-3c98-4edc-a97f-db4ee1e94307 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.532554] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for the task: (returnval){ [ 735.532554] env[61987]: value = "task-1061467" [ 735.532554] env[61987]: _type = "Task" [ 735.532554] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.541334] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061467, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.669335] env[61987]: DEBUG oslo_concurrency.lockutils [None req-514a238a-5001-456b-aa46-3e91c0dca728 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "325183b9-e35b-4f9c-a271-d0fdb05288bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.848s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.686490] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061466, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.722456] env[61987]: DEBUG oslo_concurrency.lockutils [req-9abb7de3-7729-41d8-96c6-c005a688e08d req-1ec53e81-447d-4a1f-87ae-3152204c4f58 service nova] Releasing lock "refresh_cache-ad40937f-d5a9-4708-8b81-06087d38d765" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.834036] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-45f57ee2-5fb5-4fba-86b3-0f2412b40cbf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.846704] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13fec2ff-e8e5-4a84-a865-b242276194e6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.887087] env[61987]: DEBUG nova.compute.manager [req-1e5e54d5-75b2-4927-9b4c-d6ad98f3a5ff req-4ad9c6e0-3523-42af-aecb-4d2b66ce913b service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Detach interface failed, port_id=a4965cbd-fb11-4afb-acec-bf0128f32e0f, reason: Instance 93174a10-f8b1-4789-ab3c-dda07bdddaa8 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 735.887087] env[61987]: DEBUG nova.compute.manager [req-1e5e54d5-75b2-4927-9b4c-d6ad98f3a5ff req-4ad9c6e0-3523-42af-aecb-4d2b66ce913b service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Received event network-vif-deleted-b3fafafe-77c6-41b9-ae56-b17b2d781bca {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 735.887087] env[61987]: INFO nova.compute.manager [req-1e5e54d5-75b2-4927-9b4c-d6ad98f3a5ff req-4ad9c6e0-3523-42af-aecb-4d2b66ce913b service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Neutron deleted interface b3fafafe-77c6-41b9-ae56-b17b2d781bca; detaching it from the instance and deleting it from the info cache [ 735.887087] env[61987]: DEBUG nova.network.neutron [req-1e5e54d5-75b2-4927-9b4c-d6ad98f3a5ff req-4ad9c6e0-3523-42af-aecb-4d2b66ce913b service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.923144] env[61987]: DEBUG nova.scheduler.client.report [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 735.985178] env[61987]: DEBUG nova.compute.manager [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 735.985178] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 736.777595] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf41388c-99ed-4d66-9766-a51553d47b34 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.785032] env[61987]: INFO nova.compute.manager [-] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Took 2.33 seconds to deallocate network for instance. [ 736.785456] env[61987]: DEBUG nova.compute.manager [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 736.788615] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.950s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.788965] env[61987]: DEBUG nova.compute.manager [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 736.799727] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a1edbc8c-7c30-421d-8c27-8f1dc130492a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.803475] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.344s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.803475] env[61987]: DEBUG nova.objects.instance [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Lazy-loading 'resources' on Instance uuid fa943137-caf1-4b41-b644-c9146f72f956 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 736.811318] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "325183b9-e35b-4f9c-a271-d0fdb05288bb" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.811318] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "325183b9-e35b-4f9c-a271-d0fdb05288bb" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.819137] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 736.819670] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]527d3612-8d77-c427-bcb0-55b2a0bab0fa, 'name': SearchDatastore_Task, 'duration_secs': 0.028951} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.822028] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5c2449e7-021e-4546-b9d2-0d2c279fc856 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.828028] env[61987]: DEBUG oslo_concurrency.lockutils [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.828028] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] ad40937f-d5a9-4708-8b81-06087d38d765/ad40937f-d5a9-4708-8b81-06087d38d765.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 736.830543] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-33f51cee-637a-4858-859c-f7b2666f79fc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.832870] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061467, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.562866} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.833116] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061466, 'name': ReconfigVM_Task, 'duration_secs': 0.589063} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.835417] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 1afcffda-0bd2-4cbe-8cfb-12a91bb50975/1afcffda-0bd2-4cbe-8cfb-12a91bb50975.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 736.835731] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 736.835945] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Reconfigured VM instance instance-00000030 to attach disk [datastore2] d038391c-2b32-455c-85d5-68221e7bd065/d038391c-2b32-455c-85d5-68221e7bd065.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 736.837281] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a346a10e-fd14-436e-8a39-2b8ae3aa3813 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.839083] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b6aecf27-1b25-4994-b887-fd4c7441a8b4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.843792] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ec87a50-0f97-48ce-88aa-6cd720794f32 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.861967] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 736.861967] env[61987]: value = "task-1061469" [ 736.861967] env[61987]: _type = "Task" [ 736.861967] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.863880] env[61987]: DEBUG oslo_vmware.api [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Waiting for the task: (returnval){ [ 736.863880] env[61987]: value = "task-1061468" [ 736.863880] env[61987]: _type = "Task" [ 736.863880] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.875273] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for the task: (returnval){ [ 736.875273] env[61987]: value = "task-1061470" [ 736.875273] env[61987]: _type = "Task" [ 736.875273] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.875932] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 736.875932] env[61987]: value = "task-1061471" [ 736.875932] env[61987]: _type = "Task" [ 736.875932] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.886540] env[61987]: DEBUG nova.compute.manager [req-1e5e54d5-75b2-4927-9b4c-d6ad98f3a5ff req-4ad9c6e0-3523-42af-aecb-4d2b66ce913b service nova] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Detach interface failed, port_id=b3fafafe-77c6-41b9-ae56-b17b2d781bca, reason: Instance 93174a10-f8b1-4789-ab3c-dda07bdddaa8 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 736.904621] env[61987]: DEBUG oslo_vmware.api [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': task-1061468, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.904621] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061469, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.904782] env[61987]: DEBUG oslo_vmware.rw_handles [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52899531-02de-3c63-adb0-f79a7ec521a7/disk-0.vmdk. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 736.906456] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ef80b5-61c0-437a-a7cc-a84bb44dbc3a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.916207] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061470, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.916605] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061471, 'name': Rename_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.920506] env[61987]: DEBUG oslo_vmware.rw_handles [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52899531-02de-3c63-adb0-f79a7ec521a7/disk-0.vmdk is in state: ready. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 736.920747] env[61987]: ERROR oslo_vmware.rw_handles [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52899531-02de-3c63-adb0-f79a7ec521a7/disk-0.vmdk due to incomplete transfer. [ 736.921019] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ebc6ec28-2d4b-4afe-b209-af225200b26b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.927822] env[61987]: DEBUG oslo_vmware.rw_handles [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52899531-02de-3c63-adb0-f79a7ec521a7/disk-0.vmdk. {{(pid=61987) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 736.928048] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Uploaded image 0098a037-893f-46c6-816c-12e3250d8b8d to the Glance image server {{(pid=61987) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 736.930264] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Destroying the VM {{(pid=61987) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 736.930507] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-790a0228-d485-43d0-bad2-7427af303ee2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.937624] env[61987]: DEBUG oslo_vmware.api [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for the task: (returnval){ [ 736.937624] env[61987]: value = "task-1061472" [ 736.937624] env[61987]: _type = "Task" [ 736.937624] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.948021] env[61987]: DEBUG oslo_vmware.api [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061472, 'name': Destroy_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.306474] env[61987]: DEBUG nova.compute.utils [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 737.308085] env[61987]: DEBUG nova.compute.manager [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 737.308264] env[61987]: DEBUG nova.network.neutron [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 737.322143] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.323513] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.324258] env[61987]: DEBUG nova.compute.utils [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 737.382810] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061469, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.383293] env[61987]: DEBUG oslo_vmware.api [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': task-1061468, 'name': PowerOffVM_Task, 'duration_secs': 0.363853} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.383615] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 737.383894] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 737.384236] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0d1d5952-c2c2-4a34-85b8-9cec8cec7100 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.400458] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061470, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074538} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.404053] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 737.404407] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061471, 'name': Rename_Task, 'duration_secs': 0.171478} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.405250] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82208b14-34b0-4e40-bd1c-10521768d304 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.409336] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 737.410950] env[61987]: DEBUG nova.policy [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1af6c660a24f41f9b7417a056fe6856f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0e6fb7bd076b41838cf2a873b3e4b8a0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 737.415049] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0da0fdb6-6e6a-444f-a1df-b95cb4f43307 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.438526] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 1afcffda-0bd2-4cbe-8cfb-12a91bb50975/1afcffda-0bd2-4cbe-8cfb-12a91bb50975.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 737.443912] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fefaf41a-2977-435a-84f4-ee280978c79d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.462146] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 737.462146] env[61987]: value = "task-1061474" [ 737.462146] env[61987]: _type = "Task" [ 737.462146] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.473742] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 737.474119] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 737.474388] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Deleting the datastore file [datastore2] d55395cc-0b92-4e99-9b80-913c19f6c7f4 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 737.475514] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-227ad169-ee8c-4b70-90be-a1705ef13bfa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.477776] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for the task: (returnval){ [ 737.477776] env[61987]: value = "task-1061475" [ 737.477776] env[61987]: _type = "Task" [ 737.477776] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.485106] env[61987]: DEBUG oslo_vmware.api [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061472, 'name': Destroy_Task} progress is 33%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.485615] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061474, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.493735] env[61987]: DEBUG oslo_vmware.api [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Waiting for the task: (returnval){ [ 737.493735] env[61987]: value = "task-1061476" [ 737.493735] env[61987]: _type = "Task" [ 737.493735] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.497269] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061475, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.506650] env[61987]: DEBUG oslo_vmware.api [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': task-1061476, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.815020] env[61987]: DEBUG nova.compute.manager [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 737.818594] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02bd0057-b7b9-4b73-9daa-ee6b72e5a949 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.827291] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "325183b9-e35b-4f9c-a271-d0fdb05288bb" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.017s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.828832] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-476ad40a-c0c4-4031-bb01-7ca1638781f9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.874578] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a8f016f-bf80-4f27-8b59-68a195b31c94 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.887371] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb2e9f41-9440-40af-9312-32a132577f34 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.891470] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061469, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.555382} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.891774] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] ad40937f-d5a9-4708-8b81-06087d38d765/ad40937f-d5a9-4708-8b81-06087d38d765.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 737.891997] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 737.892644] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a2addab9-6197-4dba-a7d7-5aa1a7541d5a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.903301] env[61987]: DEBUG nova.compute.provider_tree [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.906224] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 737.906224] env[61987]: value = "task-1061477" [ 737.906224] env[61987]: _type = "Task" [ 737.906224] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.916812] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061477, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.958989] env[61987]: DEBUG oslo_vmware.api [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061472, 'name': Destroy_Task, 'duration_secs': 0.689558} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.958989] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Destroyed the VM [ 737.958989] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Deleting Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 737.958989] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-41c91495-bb04-4be7-bccf-aa3fd829893d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.965634] env[61987]: DEBUG oslo_vmware.api [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for the task: (returnval){ [ 737.965634] env[61987]: value = "task-1061478" [ 737.965634] env[61987]: _type = "Task" [ 737.965634] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.973051] env[61987]: DEBUG nova.network.neutron [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Successfully created port: 021ce2e6-14d2-4f08-b654-6770827da9ee {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 737.978907] env[61987]: DEBUG oslo_vmware.api [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061478, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.984342] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061474, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.994349] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061475, 'name': ReconfigVM_Task, 'duration_secs': 0.319061} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.994560] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 1afcffda-0bd2-4cbe-8cfb-12a91bb50975/1afcffda-0bd2-4cbe-8cfb-12a91bb50975.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 737.995228] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-764d960a-d84d-457e-9661-a9682f4c8459 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.003236] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for the task: (returnval){ [ 738.003236] env[61987]: value = "task-1061479" [ 738.003236] env[61987]: _type = "Task" [ 738.003236] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.009798] env[61987]: DEBUG oslo_vmware.api [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Task: {'id': task-1061476, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169914} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.010238] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 738.010495] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 738.010730] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 738.010960] env[61987]: INFO nova.compute.manager [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Took 2.03 seconds to destroy the instance on the hypervisor. [ 738.011559] env[61987]: DEBUG oslo.service.loopingcall [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 738.011559] env[61987]: DEBUG nova.compute.manager [-] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 738.011713] env[61987]: DEBUG nova.network.neutron [-] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 738.016829] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061479, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.326344] env[61987]: INFO nova.virt.block_device [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Booting with volume a8d52043-effe-4e4a-9a52-3ed5b5050a94 at /dev/sda [ 738.374613] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9ce71620-620a-4a77-9c26-ee1d22b338e7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.383776] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6a0274-4816-4a59-ba74-33aece0a794b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.408367] env[61987]: DEBUG nova.scheduler.client.report [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 738.426962] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f1a3c4c9-4b12-4276-b005-e1d261dc0389 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.435276] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061477, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.126949} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.437479] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 738.439211] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6412193e-9a83-424d-9c97-c44250831cb8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.451428] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d040602-21af-4828-9989-72433036a34c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.511294] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] ad40937f-d5a9-4708-8b81-06087d38d765/ad40937f-d5a9-4708-8b81-06087d38d765.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 738.526277] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-788d3086-8d1a-4fc9-8983-0cc7bd306606 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.560398] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-994203c3-84c8-4cc5-a09a-461c15d530bd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.563120] env[61987]: DEBUG oslo_vmware.api [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061474, 'name': PowerOnVM_Task, 'duration_secs': 0.844779} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.563376] env[61987]: DEBUG oslo_vmware.api [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061478, 'name': RemoveSnapshot_Task} progress is 26%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.564067] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 738.564280] env[61987]: INFO nova.compute.manager [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Took 10.10 seconds to spawn the instance on the hypervisor. [ 738.564475] env[61987]: DEBUG nova.compute.manager [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 738.565816] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f2fcbf0-a5f2-41f3-99cb-53a7cc896035 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.572938] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061479, 'name': Rename_Task, 'duration_secs': 0.160518} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.575077] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13120edd-4f22-4839-95e3-a276e08e9e10 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.580185] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 738.580185] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 738.580185] env[61987]: value = "task-1061480" [ 738.580185] env[61987]: _type = "Task" [ 738.580185] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.580185] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff50218d-635d-421b-bc49-04bcaa89fc2a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.594348] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061480, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.595832] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for the task: (returnval){ [ 738.595832] env[61987]: value = "task-1061481" [ 738.595832] env[61987]: _type = "Task" [ 738.595832] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.604952] env[61987]: DEBUG nova.virt.block_device [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Updating existing volume attachment record: 74c610ea-9990-4c11-809a-138f1e416727 {{(pid=61987) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 738.612338] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061481, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.645387] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 738.645617] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 738.732805] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "325183b9-e35b-4f9c-a271-d0fdb05288bb" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.732805] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "325183b9-e35b-4f9c-a271-d0fdb05288bb" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.732805] env[61987]: INFO nova.compute.manager [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Attaching volume fcdd03f8-ce63-4537-a4ea-aed9ed31ca09 to /dev/sdb [ 738.772854] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3403ec0-4f74-49ec-8897-97a9d6bb9014 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.781152] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1da537dc-e0d8-4518-a729-5e96bbb74932 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.804442] env[61987]: DEBUG nova.virt.block_device [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Updating existing volume attachment record: 37cdd063-56ad-4ec2-8e0c-5be9022c5f69 {{(pid=61987) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 738.926035] env[61987]: DEBUG nova.compute.manager [req-9458fdc6-4f8d-42c9-a638-dbb2cf73a6d0 req-2d94058f-dcf4-4fcc-a052-8dddcf9792c2 service nova] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Received event network-vif-deleted-27b82c05-3eb6-48ab-9455-b321ef1f1fe6 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 738.926280] env[61987]: INFO nova.compute.manager [req-9458fdc6-4f8d-42c9-a638-dbb2cf73a6d0 req-2d94058f-dcf4-4fcc-a052-8dddcf9792c2 service nova] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Neutron deleted interface 27b82c05-3eb6-48ab-9455-b321ef1f1fe6; detaching it from the instance and deleting it from the info cache [ 738.926280] env[61987]: DEBUG nova.network.neutron [req-9458fdc6-4f8d-42c9-a638-dbb2cf73a6d0 req-2d94058f-dcf4-4fcc-a052-8dddcf9792c2 service nova] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.928394] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.127s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.931364] env[61987]: DEBUG oslo_concurrency.lockutils [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.081s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.931801] env[61987]: DEBUG nova.objects.instance [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Lazy-loading 'resources' on Instance uuid 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 738.953723] env[61987]: INFO nova.scheduler.client.report [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Deleted allocations for instance fa943137-caf1-4b41-b644-c9146f72f956 [ 738.982600] env[61987]: DEBUG oslo_vmware.api [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061478, 'name': RemoveSnapshot_Task, 'duration_secs': 0.612725} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.982941] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Deleted Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 738.983220] env[61987]: INFO nova.compute.manager [None req-3b47060b-e039-4fc3-bed4-59bc2aaae9df tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Took 17.84 seconds to snapshot the instance on the hypervisor. [ 739.106120] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061480, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.113583] env[61987]: INFO nova.compute.manager [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Took 27.62 seconds to build instance. [ 739.124101] env[61987]: DEBUG oslo_vmware.api [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061481, 'name': PowerOnVM_Task, 'duration_secs': 0.50939} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.124101] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 739.124101] env[61987]: INFO nova.compute.manager [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Took 5.36 seconds to spawn the instance on the hypervisor. [ 739.124101] env[61987]: DEBUG nova.compute.manager [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 739.125517] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6373d296-db51-4bb1-9085-a9a4a5fee44a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.156232] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 739.156402] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Starting heal instance info cache {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10401}} [ 739.156502] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Rebuilding the list of instances to heal {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10405}} [ 739.170396] env[61987]: DEBUG nova.network.neutron [-] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.433238] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-23ea177d-81f1-4f09-bc1e-1efb999bb67e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.451200] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30524b11-e88c-42f9-9b42-bc3ffd4369fd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.470618] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6118559-c4b7-4d3b-9663-a6e086194b60 tempest-ServerDiagnosticsTest-1900374119 tempest-ServerDiagnosticsTest-1900374119-project-member] Lock "fa943137-caf1-4b41-b644-c9146f72f956" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.208s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.498501] env[61987]: DEBUG nova.compute.manager [req-9458fdc6-4f8d-42c9-a638-dbb2cf73a6d0 req-2d94058f-dcf4-4fcc-a052-8dddcf9792c2 service nova] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Detach interface failed, port_id=27b82c05-3eb6-48ab-9455-b321ef1f1fe6, reason: Instance d55395cc-0b92-4e99-9b80-913c19f6c7f4 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 739.601967] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061480, 'name': ReconfigVM_Task, 'duration_secs': 0.728929} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.601967] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Reconfigured VM instance instance-00000031 to attach disk [datastore2] ad40937f-d5a9-4708-8b81-06087d38d765/ad40937f-d5a9-4708-8b81-06087d38d765.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 739.601967] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6b60c06c-d394-4a16-a176-16122df74e21 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.612408] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 739.612408] env[61987]: value = "task-1061485" [ 739.612408] env[61987]: _type = "Task" [ 739.612408] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.618635] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b494d12-6926-4009-90fc-85c74221d160 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Lock "d038391c-2b32-455c-85d5-68221e7bd065" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.964s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.626630] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061485, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.651361] env[61987]: INFO nova.compute.manager [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Took 21.91 seconds to build instance. [ 739.663505] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Skipping network cache update for instance because it is being deleted. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10418}} [ 739.664255] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Skipping network cache update for instance because it is Building. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10414}} [ 739.664255] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Skipping network cache update for instance because it is Building. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10414}} [ 739.664255] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Skipping network cache update for instance because it is Building. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10414}} [ 739.678556] env[61987]: INFO nova.compute.manager [-] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Took 1.67 seconds to deallocate network for instance. [ 739.763850] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "refresh_cache-630e0959-6d7f-401d-b84c-1c098f54bf60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.764114] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquired lock "refresh_cache-630e0959-6d7f-401d-b84c-1c098f54bf60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.764313] env[61987]: DEBUG nova.network.neutron [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Forcefully refreshing network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 739.764481] env[61987]: DEBUG nova.objects.instance [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lazy-loading 'info_cache' on Instance uuid 630e0959-6d7f-401d-b84c-1c098f54bf60 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 740.014569] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246de8dd-9ba7-4457-856c-490ac2333fdf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.025279] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0768a969-547e-490c-84ed-b75974ef7a16 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.062336] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ffe1eb-67e8-450f-9968-5ed485b0248c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.072261] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-425ee779-81c5-405b-a23a-2851ad2fb855 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.087458] env[61987]: DEBUG nova.compute.provider_tree [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.125908] env[61987]: DEBUG nova.compute.manager [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 740.125908] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061485, 'name': Rename_Task, 'duration_secs': 0.212629} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.125908] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 740.125908] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-88b0f02e-8c0e-45ce-8a31-151e55c066a7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.135159] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 740.135159] env[61987]: value = "task-1061486" [ 740.135159] env[61987]: _type = "Task" [ 740.135159] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.145661] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061486, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.152321] env[61987]: DEBUG oslo_concurrency.lockutils [None req-30ac1ec4-9679-49e8-acf8-362a1193f365 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Lock "1afcffda-0bd2-4cbe-8cfb-12a91bb50975" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.003s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.194267] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.252009] env[61987]: DEBUG nova.compute.manager [req-70e8fde2-315b-4a2e-983a-26cd2e66480d req-1087915e-97e0-4646-b58c-bffb5f1d894e service nova] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Received event network-vif-plugged-021ce2e6-14d2-4f08-b654-6770827da9ee {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 740.252464] env[61987]: DEBUG oslo_concurrency.lockutils [req-70e8fde2-315b-4a2e-983a-26cd2e66480d req-1087915e-97e0-4646-b58c-bffb5f1d894e service nova] Acquiring lock "95641d0b-970c-4b94-8568-5c46d0808345-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.252839] env[61987]: DEBUG oslo_concurrency.lockutils [req-70e8fde2-315b-4a2e-983a-26cd2e66480d req-1087915e-97e0-4646-b58c-bffb5f1d894e service nova] Lock "95641d0b-970c-4b94-8568-5c46d0808345-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.253149] env[61987]: DEBUG oslo_concurrency.lockutils [req-70e8fde2-315b-4a2e-983a-26cd2e66480d req-1087915e-97e0-4646-b58c-bffb5f1d894e service nova] Lock "95641d0b-970c-4b94-8568-5c46d0808345-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.253403] env[61987]: DEBUG nova.compute.manager [req-70e8fde2-315b-4a2e-983a-26cd2e66480d req-1087915e-97e0-4646-b58c-bffb5f1d894e service nova] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] No waiting events found dispatching network-vif-plugged-021ce2e6-14d2-4f08-b654-6770827da9ee {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 740.253716] env[61987]: WARNING nova.compute.manager [req-70e8fde2-315b-4a2e-983a-26cd2e66480d req-1087915e-97e0-4646-b58c-bffb5f1d894e service nova] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Received unexpected event network-vif-plugged-021ce2e6-14d2-4f08-b654-6770827da9ee for instance with vm_state building and task_state block_device_mapping. [ 740.274889] env[61987]: INFO nova.compute.manager [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Rebuilding instance [ 740.328503] env[61987]: DEBUG nova.compute.manager [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 740.331844] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-635ca101-20a8-4690-a588-61d863382097 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.382722] env[61987]: DEBUG nova.network.neutron [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Successfully updated port: 021ce2e6-14d2-4f08-b654-6770827da9ee {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 740.591422] env[61987]: DEBUG nova.scheduler.client.report [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 740.645806] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061486, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.649581] env[61987]: DEBUG oslo_concurrency.lockutils [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.655437] env[61987]: DEBUG nova.compute.manager [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 740.774957] env[61987]: DEBUG nova.compute.manager [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 740.775609] env[61987]: DEBUG nova.virt.hardware [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 740.775752] env[61987]: DEBUG nova.virt.hardware [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 740.775916] env[61987]: DEBUG nova.virt.hardware [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 740.776132] env[61987]: DEBUG nova.virt.hardware [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 740.776293] env[61987]: DEBUG nova.virt.hardware [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 740.776448] env[61987]: DEBUG nova.virt.hardware [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 740.776659] env[61987]: DEBUG nova.virt.hardware [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 740.776855] env[61987]: DEBUG nova.virt.hardware [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 740.777242] env[61987]: DEBUG nova.virt.hardware [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 740.777408] env[61987]: DEBUG nova.virt.hardware [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 740.777593] env[61987]: DEBUG nova.virt.hardware [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 740.780808] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32540e94-9e2f-407c-aa17-48b06faa69d5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.790252] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd53ad0-4e34-4fe1-b92e-9c1be9153e4d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.888757] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Acquiring lock "refresh_cache-95641d0b-970c-4b94-8568-5c46d0808345" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.888819] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Acquired lock "refresh_cache-95641d0b-970c-4b94-8568-5c46d0808345" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.888994] env[61987]: DEBUG nova.network.neutron [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 741.102089] env[61987]: DEBUG oslo_concurrency.lockutils [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.170s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.105696] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.881s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.105907] env[61987]: DEBUG nova.objects.instance [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lazy-loading 'resources' on Instance uuid ea6c2120-a69a-42c7-bc92-3e74539c008d {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 741.132621] env[61987]: INFO nova.scheduler.client.report [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Deleted allocations for instance 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f [ 741.152488] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061486, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.184882] env[61987]: DEBUG oslo_concurrency.lockutils [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.346498] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 741.347240] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-238b326a-d02c-4ff7-ad01-d72f94dc895b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.355631] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for the task: (returnval){ [ 741.355631] env[61987]: value = "task-1061488" [ 741.355631] env[61987]: _type = "Task" [ 741.355631] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.372146] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061488, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.442376] env[61987]: DEBUG nova.network.neutron [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 741.532033] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Acquiring lock "630e0959-6d7f-401d-b84c-1c098f54bf60" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.533628] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Lock "630e0959-6d7f-401d-b84c-1c098f54bf60" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.537348] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Acquiring lock "630e0959-6d7f-401d-b84c-1c098f54bf60-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.537348] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Lock "630e0959-6d7f-401d-b84c-1c098f54bf60-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.537348] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Lock "630e0959-6d7f-401d-b84c-1c098f54bf60-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.537348] env[61987]: INFO nova.compute.manager [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Terminating instance [ 741.655500] env[61987]: DEBUG oslo_concurrency.lockutils [None req-09d95990-c4a5-4442-9a60-71c61caef1d0 tempest-ServerGroupTestJSON-659753702 tempest-ServerGroupTestJSON-659753702-project-member] Lock "49e0d752-6f88-45c4-bd6a-d9fe4f526a2f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.293s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.661040] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061486, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.791691] env[61987]: DEBUG nova.network.neutron [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Updating instance_info_cache with network_info: [{"id": "0c62bc7d-9ca1-42c9-b59b-53a18c465220", "address": "fa:16:3e:71:6e:ef", "network": {"id": "c3010fca-d3b0-4624-b84b-b0920f0a0e05", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1462858397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04f95c6e34db49c58e3827d7a72cab54", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c62bc7d-9c", "ovs_interfaceid": "0c62bc7d-9ca1-42c9-b59b-53a18c465220", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.800932] env[61987]: DEBUG nova.network.neutron [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Updating instance_info_cache with network_info: [{"id": "021ce2e6-14d2-4f08-b654-6770827da9ee", "address": "fa:16:3e:96:d4:75", "network": {"id": "801ddda8-4f5d-42fe-9660-5b294986ecce", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-764406404-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e6fb7bd076b41838cf2a873b3e4b8a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe20ef0e-0991-44d7-887d-08dddac0b56b", "external-id": "nsx-vlan-transportzone-991", "segmentation_id": 991, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap021ce2e6-14", "ovs_interfaceid": "021ce2e6-14d2-4f08-b654-6770827da9ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.870618] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061488, 'name': PowerOffVM_Task, 'duration_secs': 0.159031} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.870927] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 741.871189] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 741.872024] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b56166f-3f53-4e7f-964b-1c1bbdd20978 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.886839] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 741.886839] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7cb40313-0349-4f65-941b-748bded74adf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.923725] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 741.923990] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 741.924238] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Deleting the datastore file [datastore1] 1afcffda-0bd2-4cbe-8cfb-12a91bb50975 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 741.924520] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dc7c969f-0a5d-4684-93b7-eb6bde2b0389 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.935807] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for the task: (returnval){ [ 741.935807] env[61987]: value = "task-1061490" [ 741.935807] env[61987]: _type = "Task" [ 741.935807] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.953740] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061490, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.046678] env[61987]: DEBUG nova.compute.manager [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 742.046678] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 742.047716] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feff4606-a49c-4e29-8b06-61f44c25f263 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.057646] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 742.060712] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-38d95a97-0e6d-435d-92ea-6b35c1f54b9f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.070879] env[61987]: DEBUG oslo_vmware.api [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for the task: (returnval){ [ 742.070879] env[61987]: value = "task-1061491" [ 742.070879] env[61987]: _type = "Task" [ 742.070879] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.079259] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db6d46f-c546-4ffa-9cfa-2b656de7b3f7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.091315] env[61987]: DEBUG oslo_vmware.api [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061491, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.095731] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84585748-1e68-49b9-97f5-83466218dff5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.135135] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33a613ca-e576-4ad4-b5db-d4f4e5d286fc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.147541] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2a366dc-657c-4993-92ed-aaebfc6fd832 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.158595] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061486, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.169985] env[61987]: DEBUG nova.compute.provider_tree [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.294805] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Releasing lock "refresh_cache-630e0959-6d7f-401d-b84c-1c098f54bf60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.295047] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Updated the network info_cache for instance {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10472}} [ 742.295843] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 742.295843] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 742.295843] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 742.295843] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 742.295843] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 742.296154] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 742.296154] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61987) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11020}} [ 742.296266] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 742.304558] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Releasing lock "refresh_cache-95641d0b-970c-4b94-8568-5c46d0808345" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.304558] env[61987]: DEBUG nova.compute.manager [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Instance network_info: |[{"id": "021ce2e6-14d2-4f08-b654-6770827da9ee", "address": "fa:16:3e:96:d4:75", "network": {"id": "801ddda8-4f5d-42fe-9660-5b294986ecce", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-764406404-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e6fb7bd076b41838cf2a873b3e4b8a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe20ef0e-0991-44d7-887d-08dddac0b56b", "external-id": "nsx-vlan-transportzone-991", "segmentation_id": 991, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap021ce2e6-14", "ovs_interfaceid": "021ce2e6-14d2-4f08-b654-6770827da9ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 742.304558] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:96:d4:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fe20ef0e-0991-44d7-887d-08dddac0b56b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '021ce2e6-14d2-4f08-b654-6770827da9ee', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 742.311469] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Creating folder: Project (0e6fb7bd076b41838cf2a873b3e4b8a0). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 742.311891] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bbd73543-5cb1-48a7-8e45-d62189627a00 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.328456] env[61987]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 742.328918] env[61987]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61987) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 742.329650] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Folder already exists: Project (0e6fb7bd076b41838cf2a873b3e4b8a0). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 742.331988] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Creating folder: Instances. Parent ref: group-v234226. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 742.331988] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e8c82893-4ba4-4e45-a90f-af5db09a4842 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.344812] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Created folder: Instances in parent group-v234226. [ 742.344997] env[61987]: DEBUG oslo.service.loopingcall [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 742.345340] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 742.349086] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d4b09180-c40a-4b28-adcf-cd4c48fe38ed {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.376520] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 742.376520] env[61987]: value = "task-1061494" [ 742.376520] env[61987]: _type = "Task" [ 742.376520] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.390350] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061494, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.449698] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061490, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159497} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.449997] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 742.450215] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 742.450400] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 742.471023] env[61987]: DEBUG nova.compute.manager [req-41c49f50-ff57-4abe-b295-2f9940614173 req-5d50d2d3-f240-49c3-a1f1-fdda3375c2f1 service nova] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Received event network-changed-021ce2e6-14d2-4f08-b654-6770827da9ee {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 742.471023] env[61987]: DEBUG nova.compute.manager [req-41c49f50-ff57-4abe-b295-2f9940614173 req-5d50d2d3-f240-49c3-a1f1-fdda3375c2f1 service nova] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Refreshing instance network info cache due to event network-changed-021ce2e6-14d2-4f08-b654-6770827da9ee. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 742.471023] env[61987]: DEBUG oslo_concurrency.lockutils [req-41c49f50-ff57-4abe-b295-2f9940614173 req-5d50d2d3-f240-49c3-a1f1-fdda3375c2f1 service nova] Acquiring lock "refresh_cache-95641d0b-970c-4b94-8568-5c46d0808345" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.471023] env[61987]: DEBUG oslo_concurrency.lockutils [req-41c49f50-ff57-4abe-b295-2f9940614173 req-5d50d2d3-f240-49c3-a1f1-fdda3375c2f1 service nova] Acquired lock "refresh_cache-95641d0b-970c-4b94-8568-5c46d0808345" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.471023] env[61987]: DEBUG nova.network.neutron [req-41c49f50-ff57-4abe-b295-2f9940614173 req-5d50d2d3-f240-49c3-a1f1-fdda3375c2f1 service nova] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Refreshing network info cache for port 021ce2e6-14d2-4f08-b654-6770827da9ee {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 742.585065] env[61987]: DEBUG oslo_vmware.api [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061491, 'name': PowerOffVM_Task, 'duration_secs': 0.32934} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.585761] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 742.586034] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 742.586345] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8b58b4df-9b76-4100-855f-1547bc4732d9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.653699] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 742.653985] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 742.654193] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Deleting the datastore file [datastore2] 630e0959-6d7f-401d-b84c-1c098f54bf60 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 742.655063] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c518c2be-f531-4b3a-8d7c-bb253038c147 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.659879] env[61987]: DEBUG oslo_vmware.api [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061486, 'name': PowerOnVM_Task, 'duration_secs': 2.116162} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.660690] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 742.661946] env[61987]: INFO nova.compute.manager [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Took 11.48 seconds to spawn the instance on the hypervisor. [ 742.662140] env[61987]: DEBUG nova.compute.manager [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 742.662949] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31147fab-888c-4a0d-a11b-7688c15b8800 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.666817] env[61987]: DEBUG oslo_vmware.api [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for the task: (returnval){ [ 742.666817] env[61987]: value = "task-1061496" [ 742.666817] env[61987]: _type = "Task" [ 742.666817] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.674531] env[61987]: DEBUG nova.scheduler.client.report [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 742.689266] env[61987]: DEBUG oslo_vmware.api [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061496, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.800705] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.889230] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061494, 'name': CreateVM_Task, 'duration_secs': 0.460289} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.889230] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 742.889230] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'device_type': None, 'attachment_id': '74c610ea-9990-4c11-809a-138f1e416727', 'mount_device': '/dev/sda', 'delete_on_termination': True, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234229', 'volume_id': 'a8d52043-effe-4e4a-9a52-3ed5b5050a94', 'name': 'volume-a8d52043-effe-4e4a-9a52-3ed5b5050a94', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '95641d0b-970c-4b94-8568-5c46d0808345', 'attached_at': '', 'detached_at': '', 'volume_id': 'a8d52043-effe-4e4a-9a52-3ed5b5050a94', 'serial': 'a8d52043-effe-4e4a-9a52-3ed5b5050a94'}, 'boot_index': 0, 'guest_format': None, 'disk_bus': None, 'volume_type': None}], 'swap': None} {{(pid=61987) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 742.889230] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Root volume attach. Driver type: vmdk {{(pid=61987) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 742.889230] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3c55abb-952e-4640-9380-e3e8b32ad155 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.897807] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d11ef83-f797-4553-8734-1c090f206faf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.906828] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b5960f5-e06a-41ce-8cc4-0ccef2600fa2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.913109] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-16af9525-abeb-4675-bb14-c3bfb6662712 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.921037] env[61987]: DEBUG oslo_vmware.api [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Waiting for the task: (returnval){ [ 742.921037] env[61987]: value = "task-1061497" [ 742.921037] env[61987]: _type = "Task" [ 742.921037] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.929591] env[61987]: DEBUG oslo_vmware.api [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Task: {'id': task-1061497, 'name': RelocateVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.074291] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "84a4a69b-04db-4546-ba89-9039e382a0c4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.074725] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "84a4a69b-04db-4546-ba89-9039e382a0c4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.185652] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.080s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.196111] env[61987]: DEBUG oslo_vmware.api [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Task: {'id': task-1061496, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.381779} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.199907] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.879s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.203195] env[61987]: INFO nova.compute.claims [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 743.206344] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 743.206650] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 743.206809] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 743.206956] env[61987]: INFO nova.compute.manager [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Took 1.16 seconds to destroy the instance on the hypervisor. [ 743.207251] env[61987]: DEBUG oslo.service.loopingcall [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 743.207926] env[61987]: INFO nova.compute.manager [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Took 28.17 seconds to build instance. [ 743.211587] env[61987]: DEBUG nova.compute.manager [-] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 743.211849] env[61987]: DEBUG nova.network.neutron [-] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 743.237212] env[61987]: INFO nova.scheduler.client.report [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Deleted allocations for instance ea6c2120-a69a-42c7-bc92-3e74539c008d [ 743.374932] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Volume attach. Driver type: vmdk {{(pid=61987) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 743.375364] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234299', 'volume_id': 'fcdd03f8-ce63-4537-a4ea-aed9ed31ca09', 'name': 'volume-fcdd03f8-ce63-4537-a4ea-aed9ed31ca09', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '325183b9-e35b-4f9c-a271-d0fdb05288bb', 'attached_at': '', 'detached_at': '', 'volume_id': 'fcdd03f8-ce63-4537-a4ea-aed9ed31ca09', 'serial': 'fcdd03f8-ce63-4537-a4ea-aed9ed31ca09'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 743.376936] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47d18d0b-2335-4b5c-ac1b-37d54942fb36 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.408512] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eca2c89-f3a1-40f2-bf31-44142ae2639a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.460477] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Reconfiguring VM instance instance-0000002f to attach disk [datastore2] volume-fcdd03f8-ce63-4537-a4ea-aed9ed31ca09/volume-fcdd03f8-ce63-4537-a4ea-aed9ed31ca09.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 743.470247] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-533d7f9e-c140-404f-8614-130a8479cd97 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.495943] env[61987]: DEBUG nova.network.neutron [req-41c49f50-ff57-4abe-b295-2f9940614173 req-5d50d2d3-f240-49c3-a1f1-fdda3375c2f1 service nova] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Updated VIF entry in instance network info cache for port 021ce2e6-14d2-4f08-b654-6770827da9ee. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 743.496542] env[61987]: DEBUG nova.network.neutron [req-41c49f50-ff57-4abe-b295-2f9940614173 req-5d50d2d3-f240-49c3-a1f1-fdda3375c2f1 service nova] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Updating instance_info_cache with network_info: [{"id": "021ce2e6-14d2-4f08-b654-6770827da9ee", "address": "fa:16:3e:96:d4:75", "network": {"id": "801ddda8-4f5d-42fe-9660-5b294986ecce", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-764406404-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e6fb7bd076b41838cf2a873b3e4b8a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe20ef0e-0991-44d7-887d-08dddac0b56b", "external-id": "nsx-vlan-transportzone-991", "segmentation_id": 991, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap021ce2e6-14", "ovs_interfaceid": "021ce2e6-14d2-4f08-b654-6770827da9ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.508926] env[61987]: DEBUG oslo_vmware.api [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Task: {'id': task-1061497, 'name': RelocateVM_Task, 'duration_secs': 0.543067} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.510429] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Volume attach. Driver type: vmdk {{(pid=61987) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 743.510801] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234229', 'volume_id': 'a8d52043-effe-4e4a-9a52-3ed5b5050a94', 'name': 'volume-a8d52043-effe-4e4a-9a52-3ed5b5050a94', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '95641d0b-970c-4b94-8568-5c46d0808345', 'attached_at': '', 'detached_at': '', 'volume_id': 'a8d52043-effe-4e4a-9a52-3ed5b5050a94', 'serial': 'a8d52043-effe-4e4a-9a52-3ed5b5050a94'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 743.511393] env[61987]: DEBUG oslo_vmware.api [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 743.511393] env[61987]: value = "task-1061498" [ 743.511393] env[61987]: _type = "Task" [ 743.511393] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.514875] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05f237eb-3489-4233-a7df-9a28b2b76dd9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.552046] env[61987]: DEBUG oslo_vmware.api [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061498, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.555997] env[61987]: DEBUG nova.virt.hardware [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 743.559506] env[61987]: DEBUG nova.virt.hardware [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 743.559506] env[61987]: DEBUG nova.virt.hardware [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 743.559506] env[61987]: DEBUG nova.virt.hardware [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 743.559506] env[61987]: DEBUG nova.virt.hardware [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 743.559506] env[61987]: DEBUG nova.virt.hardware [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 743.559506] env[61987]: DEBUG nova.virt.hardware [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 743.559506] env[61987]: DEBUG nova.virt.hardware [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 743.559506] env[61987]: DEBUG nova.virt.hardware [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 743.559506] env[61987]: DEBUG nova.virt.hardware [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 743.559506] env[61987]: DEBUG nova.virt.hardware [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 743.560404] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e221ce20-55a7-42d9-a96f-a316bb98c7d7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.566176] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21851a18-8d0b-4d12-adac-fb944301eb1c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.595830] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dd5a1f3-9336-4f3c-aaa4-c8a0eb5c8571 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.612916] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] volume-a8d52043-effe-4e4a-9a52-3ed5b5050a94/volume-a8d52043-effe-4e4a-9a52-3ed5b5050a94.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 743.613689] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9fb206e7-a395-4ca9-a4e6-c18c5b0bf528 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.642435] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Instance VIF info [] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 743.648206] env[61987]: DEBUG oslo.service.loopingcall [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 743.650191] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 743.650191] env[61987]: DEBUG oslo_vmware.api [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Waiting for the task: (returnval){ [ 743.650191] env[61987]: value = "task-1061499" [ 743.650191] env[61987]: _type = "Task" [ 743.650191] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.650553] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-052580ac-1ba0-4c31-a324-8bb8f60b3a1a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.671467] env[61987]: DEBUG oslo_vmware.api [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Task: {'id': task-1061499, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.674634] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 743.674634] env[61987]: value = "task-1061500" [ 743.674634] env[61987]: _type = "Task" [ 743.674634] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.684550] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061500, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.715377] env[61987]: DEBUG oslo_concurrency.lockutils [None req-94c22a48-7ad9-4166-8a7e-c7346b67c6b6 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ad40937f-d5a9-4708-8b81-06087d38d765" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.640s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.750720] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bfdd626e-a4ed-4bef-bdf9-346d70c25945 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "ea6c2120-a69a-42c7-bc92-3e74539c008d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.053s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.001147] env[61987]: DEBUG oslo_concurrency.lockutils [req-41c49f50-ff57-4abe-b295-2f9940614173 req-5d50d2d3-f240-49c3-a1f1-fdda3375c2f1 service nova] Releasing lock "refresh_cache-95641d0b-970c-4b94-8568-5c46d0808345" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.001472] env[61987]: DEBUG nova.compute.manager [req-41c49f50-ff57-4abe-b295-2f9940614173 req-5d50d2d3-f240-49c3-a1f1-fdda3375c2f1 service nova] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Received event network-changed-203d9e60-9c1f-4c51-9ee8-9a34be19d289 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 744.001791] env[61987]: DEBUG nova.compute.manager [req-41c49f50-ff57-4abe-b295-2f9940614173 req-5d50d2d3-f240-49c3-a1f1-fdda3375c2f1 service nova] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Refreshing instance network info cache due to event network-changed-203d9e60-9c1f-4c51-9ee8-9a34be19d289. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 744.005264] env[61987]: DEBUG oslo_concurrency.lockutils [req-41c49f50-ff57-4abe-b295-2f9940614173 req-5d50d2d3-f240-49c3-a1f1-fdda3375c2f1 service nova] Acquiring lock "refresh_cache-d038391c-2b32-455c-85d5-68221e7bd065" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.005264] env[61987]: DEBUG oslo_concurrency.lockutils [req-41c49f50-ff57-4abe-b295-2f9940614173 req-5d50d2d3-f240-49c3-a1f1-fdda3375c2f1 service nova] Acquired lock "refresh_cache-d038391c-2b32-455c-85d5-68221e7bd065" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.005264] env[61987]: DEBUG nova.network.neutron [req-41c49f50-ff57-4abe-b295-2f9940614173 req-5d50d2d3-f240-49c3-a1f1-fdda3375c2f1 service nova] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Refreshing network info cache for port 203d9e60-9c1f-4c51-9ee8-9a34be19d289 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 744.032675] env[61987]: DEBUG oslo_vmware.api [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061498, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.084913] env[61987]: DEBUG nova.compute.manager [req-1c731fc1-a82a-4c56-9799-58951e5ec3c5 req-084a3e9e-7a4e-4ab5-8869-4ce238b5769e service nova] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Received event network-vif-deleted-0c62bc7d-9ca1-42c9-b59b-53a18c465220 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 744.085361] env[61987]: INFO nova.compute.manager [req-1c731fc1-a82a-4c56-9799-58951e5ec3c5 req-084a3e9e-7a4e-4ab5-8869-4ce238b5769e service nova] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Neutron deleted interface 0c62bc7d-9ca1-42c9-b59b-53a18c465220; detaching it from the instance and deleting it from the info cache [ 744.085676] env[61987]: DEBUG nova.network.neutron [req-1c731fc1-a82a-4c56-9799-58951e5ec3c5 req-084a3e9e-7a4e-4ab5-8869-4ce238b5769e service nova] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.173911] env[61987]: DEBUG oslo_vmware.api [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Task: {'id': task-1061499, 'name': ReconfigVM_Task, 'duration_secs': 0.371346} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.181933] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Reconfigured VM instance instance-00000033 to attach disk [datastore2] volume-a8d52043-effe-4e4a-9a52-3ed5b5050a94/volume-a8d52043-effe-4e4a-9a52-3ed5b5050a94.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 744.187207] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64352b9a-c803-4a12-ad91-4b550e7a6316 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.207974] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061500, 'name': CreateVM_Task, 'duration_secs': 0.35986} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.209232] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 744.209633] env[61987]: DEBUG oslo_vmware.api [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Waiting for the task: (returnval){ [ 744.209633] env[61987]: value = "task-1061501" [ 744.209633] env[61987]: _type = "Task" [ 744.209633] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.210011] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.210193] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.210519] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 744.210816] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86015fe2-2e1a-47dc-98bd-6aa7d6e8613e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.218371] env[61987]: DEBUG nova.compute.manager [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 744.230317] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for the task: (returnval){ [ 744.230317] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52128bc3-ac93-1c33-b52b-8ec0ee9ad3f4" [ 744.230317] env[61987]: _type = "Task" [ 744.230317] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.230568] env[61987]: DEBUG oslo_vmware.api [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Task: {'id': task-1061501, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.245423] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52128bc3-ac93-1c33-b52b-8ec0ee9ad3f4, 'name': SearchDatastore_Task, 'duration_secs': 0.013467} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.245776] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.246040] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 744.246302] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.246456] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.246638] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 744.246913] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f9ac10f7-aac3-47bc-993d-8b5ff9e35299 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.259580] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 744.259901] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 744.260673] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfe1c29d-52a5-469e-afe8-cb022344da51 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.270698] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for the task: (returnval){ [ 744.270698] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52df6593-d6b5-f9bd-40f0-882d75013a60" [ 744.270698] env[61987]: _type = "Task" [ 744.270698] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.282797] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52df6593-d6b5-f9bd-40f0-882d75013a60, 'name': SearchDatastore_Task, 'duration_secs': 0.011325} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.286982] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa506317-6214-4feb-91cf-f784246a7d3a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.294078] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for the task: (returnval){ [ 744.294078] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]524b37a3-fcc1-c16f-a93b-578f315d18d3" [ 744.294078] env[61987]: _type = "Task" [ 744.294078] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.304616] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524b37a3-fcc1-c16f-a93b-578f315d18d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.540142] env[61987]: DEBUG oslo_vmware.api [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061498, 'name': ReconfigVM_Task, 'duration_secs': 0.530428} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.540564] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Reconfigured VM instance instance-0000002f to attach disk [datastore2] volume-fcdd03f8-ce63-4537-a4ea-aed9ed31ca09/volume-fcdd03f8-ce63-4537-a4ea-aed9ed31ca09.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 744.545734] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-892760a4-ae27-4297-bd4f-00dea1ccbe95 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.564316] env[61987]: DEBUG nova.network.neutron [-] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.574330] env[61987]: DEBUG oslo_vmware.api [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 744.574330] env[61987]: value = "task-1061502" [ 744.574330] env[61987]: _type = "Task" [ 744.574330] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.586273] env[61987]: DEBUG oslo_vmware.api [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061502, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.595220] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-613654a1-85f3-40e3-8f6e-7b0d4fe77fe3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.595867] env[61987]: DEBUG nova.compute.manager [req-3b7cea20-5a28-4bde-9d1e-9893a62ca83a req-88dd7b8f-c967-4cb4-a926-266234da5c08 service nova] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Received event network-changed-6aa03b99-d58f-461e-b2a8-c4552c1d418a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 744.598067] env[61987]: DEBUG nova.compute.manager [req-3b7cea20-5a28-4bde-9d1e-9893a62ca83a req-88dd7b8f-c967-4cb4-a926-266234da5c08 service nova] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Refreshing instance network info cache due to event network-changed-6aa03b99-d58f-461e-b2a8-c4552c1d418a. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 744.598067] env[61987]: DEBUG oslo_concurrency.lockutils [req-3b7cea20-5a28-4bde-9d1e-9893a62ca83a req-88dd7b8f-c967-4cb4-a926-266234da5c08 service nova] Acquiring lock "refresh_cache-ad40937f-d5a9-4708-8b81-06087d38d765" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.598067] env[61987]: DEBUG oslo_concurrency.lockutils [req-3b7cea20-5a28-4bde-9d1e-9893a62ca83a req-88dd7b8f-c967-4cb4-a926-266234da5c08 service nova] Acquired lock "refresh_cache-ad40937f-d5a9-4708-8b81-06087d38d765" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.598067] env[61987]: DEBUG nova.network.neutron [req-3b7cea20-5a28-4bde-9d1e-9893a62ca83a req-88dd7b8f-c967-4cb4-a926-266234da5c08 service nova] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Refreshing network info cache for port 6aa03b99-d58f-461e-b2a8-c4552c1d418a {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 744.611142] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-039e9545-6547-476d-99e9-a2ea8ea8faf6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.647116] env[61987]: DEBUG nova.compute.manager [req-1c731fc1-a82a-4c56-9799-58951e5ec3c5 req-084a3e9e-7a4e-4ab5-8869-4ce238b5769e service nova] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Detach interface failed, port_id=0c62bc7d-9ca1-42c9-b59b-53a18c465220, reason: Instance 630e0959-6d7f-401d-b84c-1c098f54bf60 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 744.725417] env[61987]: DEBUG oslo_vmware.api [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Task: {'id': task-1061501, 'name': ReconfigVM_Task, 'duration_secs': 0.253493} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.728219] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234229', 'volume_id': 'a8d52043-effe-4e4a-9a52-3ed5b5050a94', 'name': 'volume-a8d52043-effe-4e4a-9a52-3ed5b5050a94', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '95641d0b-970c-4b94-8568-5c46d0808345', 'attached_at': '', 'detached_at': '', 'volume_id': 'a8d52043-effe-4e4a-9a52-3ed5b5050a94', 'serial': 'a8d52043-effe-4e4a-9a52-3ed5b5050a94'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 744.733050] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-29d436ea-6649-4665-a230-1e471793c0ff {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.743911] env[61987]: DEBUG oslo_vmware.api [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Waiting for the task: (returnval){ [ 744.743911] env[61987]: value = "task-1061503" [ 744.743911] env[61987]: _type = "Task" [ 744.743911] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.753174] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f8c86e1-5981-4b5d-948e-4371bd644c75 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.760024] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.760992] env[61987]: DEBUG oslo_vmware.api [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Task: {'id': task-1061503, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.766848] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35752bad-65a9-47f4-b3ae-dc8a31aae704 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.813744] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f91966b1-62dc-4d54-94b1-ddad5edc71d4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.826524] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5306e9-111b-4596-b957-16b2169b46bd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.830377] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524b37a3-fcc1-c16f-a93b-578f315d18d3, 'name': SearchDatastore_Task, 'duration_secs': 0.017926} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.832017] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.832017] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 1afcffda-0bd2-4cbe-8cfb-12a91bb50975/1afcffda-0bd2-4cbe-8cfb-12a91bb50975.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 744.832017] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5349a384-87b5-42d6-9143-e7b2b0c301b1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.842325] env[61987]: DEBUG nova.compute.provider_tree [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.850220] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for the task: (returnval){ [ 744.850220] env[61987]: value = "task-1061504" [ 744.850220] env[61987]: _type = "Task" [ 744.850220] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.859456] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061504, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.066831] env[61987]: INFO nova.compute.manager [-] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Took 1.85 seconds to deallocate network for instance. [ 745.095059] env[61987]: DEBUG oslo_vmware.api [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061502, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.095059] env[61987]: DEBUG nova.network.neutron [req-41c49f50-ff57-4abe-b295-2f9940614173 req-5d50d2d3-f240-49c3-a1f1-fdda3375c2f1 service nova] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Updated VIF entry in instance network info cache for port 203d9e60-9c1f-4c51-9ee8-9a34be19d289. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 745.095059] env[61987]: DEBUG nova.network.neutron [req-41c49f50-ff57-4abe-b295-2f9940614173 req-5d50d2d3-f240-49c3-a1f1-fdda3375c2f1 service nova] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Updating instance_info_cache with network_info: [{"id": "203d9e60-9c1f-4c51-9ee8-9a34be19d289", "address": "fa:16:3e:f1:83:58", "network": {"id": "35c53c0f-af9d-4934-84c7-ea4c5252fa58", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-343092073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.133", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d32dc833164a428dad56d1eeb9a7c3eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "790c811b-3fa6-49f8-87ac-c51450911137", "external-id": "nsx-vlan-transportzone-908", "segmentation_id": 908, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap203d9e60-9c", "ovs_interfaceid": "203d9e60-9c1f-4c51-9ee8-9a34be19d289", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.259428] env[61987]: DEBUG oslo_vmware.api [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Task: {'id': task-1061503, 'name': Rename_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.348023] env[61987]: DEBUG nova.scheduler.client.report [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 745.365757] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061504, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.492956} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.365965] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 1afcffda-0bd2-4cbe-8cfb-12a91bb50975/1afcffda-0bd2-4cbe-8cfb-12a91bb50975.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 745.366261] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 745.367133] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-597b07b2-67dc-4205-ab9f-2d959b63d793 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.377499] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for the task: (returnval){ [ 745.377499] env[61987]: value = "task-1061505" [ 745.377499] env[61987]: _type = "Task" [ 745.377499] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.387095] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061505, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.497186] env[61987]: DEBUG nova.network.neutron [req-3b7cea20-5a28-4bde-9d1e-9893a62ca83a req-88dd7b8f-c967-4cb4-a926-266234da5c08 service nova] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Updated VIF entry in instance network info cache for port 6aa03b99-d58f-461e-b2a8-c4552c1d418a. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 745.497586] env[61987]: DEBUG nova.network.neutron [req-3b7cea20-5a28-4bde-9d1e-9893a62ca83a req-88dd7b8f-c967-4cb4-a926-266234da5c08 service nova] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Updating instance_info_cache with network_info: [{"id": "6aa03b99-d58f-461e-b2a8-c4552c1d418a", "address": "fa:16:3e:5e:3e:15", "network": {"id": "a9093c43-6bc4-45b1-bdcf-a434759300bd", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-355385599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.128", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7efd6c745f547d3803d5a6184547bda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac563aa7-6d7c-4bd5-9241-7b3e11b8f22d", "external-id": "nsx-vlan-transportzone-730", "segmentation_id": 730, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6aa03b99-d5", "ovs_interfaceid": "6aa03b99-d58f-461e-b2a8-c4552c1d418a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.582082] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.592980] env[61987]: DEBUG oslo_vmware.api [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061502, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.597445] env[61987]: DEBUG oslo_concurrency.lockutils [req-41c49f50-ff57-4abe-b295-2f9940614173 req-5d50d2d3-f240-49c3-a1f1-fdda3375c2f1 service nova] Releasing lock "refresh_cache-d038391c-2b32-455c-85d5-68221e7bd065" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.756970] env[61987]: DEBUG oslo_vmware.api [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Task: {'id': task-1061503, 'name': Rename_Task, 'duration_secs': 0.836636} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.757809] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 745.758311] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9335b18c-b33b-415f-9645-04a599934a0b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.765477] env[61987]: DEBUG oslo_vmware.api [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Waiting for the task: (returnval){ [ 745.765477] env[61987]: value = "task-1061506" [ 745.765477] env[61987]: _type = "Task" [ 745.765477] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.774610] env[61987]: DEBUG oslo_vmware.api [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Task: {'id': task-1061506, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.856548] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.656s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.858371] env[61987]: DEBUG nova.compute.manager [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 745.861573] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.428s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.863522] env[61987]: INFO nova.compute.claims [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 745.888844] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061505, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074445} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.889317] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 745.890037] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-254f3bc6-e723-4f32-afe9-acde97426492 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.922716] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 1afcffda-0bd2-4cbe-8cfb-12a91bb50975/1afcffda-0bd2-4cbe-8cfb-12a91bb50975.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 745.922716] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-384aae6f-2f69-4541-8c6d-366aa5a02938 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.945296] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for the task: (returnval){ [ 745.945296] env[61987]: value = "task-1061507" [ 745.945296] env[61987]: _type = "Task" [ 745.945296] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.957272] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061507, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.002078] env[61987]: DEBUG oslo_concurrency.lockutils [req-3b7cea20-5a28-4bde-9d1e-9893a62ca83a req-88dd7b8f-c967-4cb4-a926-266234da5c08 service nova] Releasing lock "refresh_cache-ad40937f-d5a9-4708-8b81-06087d38d765" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.097523] env[61987]: DEBUG oslo_vmware.api [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061502, 'name': ReconfigVM_Task, 'duration_secs': 1.037014} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.098675] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234299', 'volume_id': 'fcdd03f8-ce63-4537-a4ea-aed9ed31ca09', 'name': 'volume-fcdd03f8-ce63-4537-a4ea-aed9ed31ca09', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '325183b9-e35b-4f9c-a271-d0fdb05288bb', 'attached_at': '', 'detached_at': '', 'volume_id': 'fcdd03f8-ce63-4537-a4ea-aed9ed31ca09', 'serial': 'fcdd03f8-ce63-4537-a4ea-aed9ed31ca09'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 746.216518] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "4fd8fc17-fb13-42e7-9879-9f120fdcb330" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.216518] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "4fd8fc17-fb13-42e7-9879-9f120fdcb330" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.279912] env[61987]: DEBUG oslo_vmware.api [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Task: {'id': task-1061506, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.371020] env[61987]: DEBUG nova.compute.utils [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 746.372444] env[61987]: DEBUG nova.compute.manager [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 746.372803] env[61987]: DEBUG nova.network.neutron [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 746.431697] env[61987]: DEBUG nova.policy [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4ead7955f1d475590523ac98c579956', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '315a2a3450024ed38e27cd7758f392f9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 746.468723] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061507, 'name': ReconfigVM_Task, 'duration_secs': 0.384401} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.469153] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 1afcffda-0bd2-4cbe-8cfb-12a91bb50975/1afcffda-0bd2-4cbe-8cfb-12a91bb50975.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 746.472108] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7c0263e5-69be-4263-989a-9a3b64a69a43 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.489048] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for the task: (returnval){ [ 746.489048] env[61987]: value = "task-1061508" [ 746.489048] env[61987]: _type = "Task" [ 746.489048] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.502427] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061508, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.778819] env[61987]: DEBUG oslo_vmware.api [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Task: {'id': task-1061506, 'name': PowerOnVM_Task, 'duration_secs': 0.98248} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.779215] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 746.779359] env[61987]: INFO nova.compute.manager [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Took 6.00 seconds to spawn the instance on the hypervisor. [ 746.779543] env[61987]: DEBUG nova.compute.manager [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 746.780355] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87cac487-5480-46f6-a33a-aae2e9873598 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.873360] env[61987]: DEBUG nova.network.neutron [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Successfully created port: 55e46e63-9076-400e-a73a-19fc4bb21bec {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 746.880359] env[61987]: DEBUG nova.compute.manager [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 747.012244] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061508, 'name': Rename_Task, 'duration_secs': 0.282214} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.012244] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 747.012244] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3f51be1c-9faf-4345-89ce-156222e51485 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.020160] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for the task: (returnval){ [ 747.020160] env[61987]: value = "task-1061509" [ 747.020160] env[61987]: _type = "Task" [ 747.020160] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.027636] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061509, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.164310] env[61987]: DEBUG nova.objects.instance [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lazy-loading 'flavor' on Instance uuid 325183b9-e35b-4f9c-a271-d0fdb05288bb {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 747.301419] env[61987]: INFO nova.compute.manager [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Took 27.55 seconds to build instance. [ 747.340902] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58bab067-bfd0-4393-9230-41e8abe34438 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.346812] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d81003fe-35ca-4626-a65d-a7b98e199f47 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.386077] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a82653c-d868-485c-b56e-027c176b59f3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.399723] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdc0d9a4-a082-4945-b46a-f11a9aa0e2fb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.413523] env[61987]: DEBUG nova.compute.provider_tree [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.532960] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061509, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.670676] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ecaa4e28-1eb9-4b28-9b7f-5ac0e1c5e275 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "325183b9-e35b-4f9c-a271-d0fdb05288bb" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.940s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.804446] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b3c220a9-9dc0-4c66-ae30-5223087ec257 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Lock "95641d0b-970c-4b94-8568-5c46d0808345" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.442s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.893890] env[61987]: DEBUG nova.compute.manager [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 747.916317] env[61987]: DEBUG nova.scheduler.client.report [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 747.921894] env[61987]: DEBUG nova.virt.hardware [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 747.922050] env[61987]: DEBUG nova.virt.hardware [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 747.922186] env[61987]: DEBUG nova.virt.hardware [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 747.922375] env[61987]: DEBUG nova.virt.hardware [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 747.922528] env[61987]: DEBUG nova.virt.hardware [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 747.922680] env[61987]: DEBUG nova.virt.hardware [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 747.922886] env[61987]: DEBUG nova.virt.hardware [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 747.923060] env[61987]: DEBUG nova.virt.hardware [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 747.923240] env[61987]: DEBUG nova.virt.hardware [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 747.923403] env[61987]: DEBUG nova.virt.hardware [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 747.923579] env[61987]: DEBUG nova.virt.hardware [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 747.924535] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b897cd3d-ca14-412e-9403-505d674cc799 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.933744] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74264c9e-45fa-431d-8d55-3137c29ab8cf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.034062] env[61987]: DEBUG oslo_vmware.api [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061509, 'name': PowerOnVM_Task, 'duration_secs': 0.638255} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.034062] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 748.035582] env[61987]: DEBUG nova.compute.manager [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 748.035582] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbe0caa5-1f28-48de-883c-4394f33a992f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.095913] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.096426] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.280653] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "325183b9-e35b-4f9c-a271-d0fdb05288bb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.281142] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "325183b9-e35b-4f9c-a271-d0fdb05288bb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.281264] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "325183b9-e35b-4f9c-a271-d0fdb05288bb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.281439] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "325183b9-e35b-4f9c-a271-d0fdb05288bb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.281524] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "325183b9-e35b-4f9c-a271-d0fdb05288bb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.284169] env[61987]: INFO nova.compute.manager [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Terminating instance [ 748.307198] env[61987]: DEBUG nova.compute.manager [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 748.429006] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.567s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.429564] env[61987]: DEBUG nova.compute.manager [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 748.433481] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 21.451s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.556627] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.788516] env[61987]: DEBUG nova.compute.manager [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 748.788770] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 748.789088] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1bd17982-e28d-44f3-b19d-387e71a8db5e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.800910] env[61987]: DEBUG oslo_vmware.api [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 748.800910] env[61987]: value = "task-1061510" [ 748.800910] env[61987]: _type = "Task" [ 748.800910] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.815481] env[61987]: DEBUG oslo_vmware.api [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061510, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.845439] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.915540] env[61987]: DEBUG nova.compute.manager [req-cbd30267-3570-4007-b62c-f9339adbd397 req-a73c412a-40c0-426c-8c56-566e6015b5fe service nova] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Received event network-vif-plugged-55e46e63-9076-400e-a73a-19fc4bb21bec {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 748.916398] env[61987]: DEBUG oslo_concurrency.lockutils [req-cbd30267-3570-4007-b62c-f9339adbd397 req-a73c412a-40c0-426c-8c56-566e6015b5fe service nova] Acquiring lock "8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.916398] env[61987]: DEBUG oslo_concurrency.lockutils [req-cbd30267-3570-4007-b62c-f9339adbd397 req-a73c412a-40c0-426c-8c56-566e6015b5fe service nova] Lock "8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.917230] env[61987]: DEBUG oslo_concurrency.lockutils [req-cbd30267-3570-4007-b62c-f9339adbd397 req-a73c412a-40c0-426c-8c56-566e6015b5fe service nova] Lock "8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.920085] env[61987]: DEBUG nova.compute.manager [req-cbd30267-3570-4007-b62c-f9339adbd397 req-a73c412a-40c0-426c-8c56-566e6015b5fe service nova] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] No waiting events found dispatching network-vif-plugged-55e46e63-9076-400e-a73a-19fc4bb21bec {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 748.920742] env[61987]: WARNING nova.compute.manager [req-cbd30267-3570-4007-b62c-f9339adbd397 req-a73c412a-40c0-426c-8c56-566e6015b5fe service nova] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Received unexpected event network-vif-plugged-55e46e63-9076-400e-a73a-19fc4bb21bec for instance with vm_state building and task_state spawning. [ 748.937143] env[61987]: DEBUG nova.compute.utils [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 748.945298] env[61987]: INFO nova.compute.claims [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 748.950949] env[61987]: DEBUG nova.compute.manager [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 748.953512] env[61987]: DEBUG nova.compute.manager [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 748.957033] env[61987]: DEBUG nova.network.neutron [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 749.019140] env[61987]: DEBUG nova.policy [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '27ae5b57556245e7b27ccc1363485352', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a53eb8ad8274fe18fd149c44eb0ffbb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 749.104822] env[61987]: DEBUG nova.network.neutron [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Successfully updated port: 55e46e63-9076-400e-a73a-19fc4bb21bec {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 749.317036] env[61987]: DEBUG oslo_vmware.api [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061510, 'name': PowerOffVM_Task, 'duration_secs': 0.314137} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.317036] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 749.317036] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Volume detach. Driver type: vmdk {{(pid=61987) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 749.317036] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234299', 'volume_id': 'fcdd03f8-ce63-4537-a4ea-aed9ed31ca09', 'name': 'volume-fcdd03f8-ce63-4537-a4ea-aed9ed31ca09', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '325183b9-e35b-4f9c-a271-d0fdb05288bb', 'attached_at': '', 'detached_at': '', 'volume_id': 'fcdd03f8-ce63-4537-a4ea-aed9ed31ca09', 'serial': 'fcdd03f8-ce63-4537-a4ea-aed9ed31ca09'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 749.317465] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6316929-ab81-452d-a4a9-2b3ed20d2a1e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.345011] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a677a333-6cae-4edb-9457-5bb239f38e7a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.355023] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d8e3310-85cf-4781-a76a-125ef1b44ea8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.380757] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b0f6a58-4baa-4adc-8ba7-08ab58cb6fe5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.404788] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] The volume has not been displaced from its original location: [datastore2] volume-fcdd03f8-ce63-4537-a4ea-aed9ed31ca09/volume-fcdd03f8-ce63-4537-a4ea-aed9ed31ca09.vmdk. No consolidation needed. {{(pid=61987) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 749.410694] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Reconfiguring VM instance instance-0000002f to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 749.411248] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f547d2fd-2bfc-408e-b8c5-7c5b3997074b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.431632] env[61987]: DEBUG oslo_vmware.api [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 749.431632] env[61987]: value = "task-1061511" [ 749.431632] env[61987]: _type = "Task" [ 749.431632] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.443171] env[61987]: DEBUG oslo_vmware.api [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061511, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.457103] env[61987]: DEBUG nova.network.neutron [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Successfully created port: 04dbba0f-63b4-4d42-a3ec-7fe8dbd70488 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 749.461294] env[61987]: INFO nova.compute.resource_tracker [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Updating resource usage from migration 0f243de7-62e3-4127-8bbb-f1c7b76ff14d [ 749.610992] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "refresh_cache-8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.612030] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired lock "refresh_cache-8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.612249] env[61987]: DEBUG nova.network.neutron [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 749.783870] env[61987]: INFO nova.compute.manager [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Rebuilding instance [ 749.831539] env[61987]: DEBUG nova.compute.manager [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 749.832461] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2fd099e-767f-4361-a6fa-00577acbe7a0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.912476] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c37a1b-4247-45d8-849f-c4a8c203c870 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.924388] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c60ed18-bed9-4245-ac24-b6a57e0dfe57 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.960203] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3253dc8-05ea-4f8f-990e-7831027ffa94 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.969289] env[61987]: DEBUG oslo_vmware.api [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061511, 'name': ReconfigVM_Task, 'duration_secs': 0.317289} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.971434] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Reconfigured VM instance instance-0000002f to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 749.977308] env[61987]: DEBUG nova.compute.manager [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 749.979704] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6faaa4f4-8037-4e48-9bcc-0ca0646f6141 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.990799] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d9dadb6-62c7-46cd-b814-3e639f8d643d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.010453] env[61987]: DEBUG nova.compute.provider_tree [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.016166] env[61987]: DEBUG nova.virt.hardware [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 750.016166] env[61987]: DEBUG nova.virt.hardware [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 750.016166] env[61987]: DEBUG nova.virt.hardware [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 750.016166] env[61987]: DEBUG nova.virt.hardware [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 750.016166] env[61987]: DEBUG nova.virt.hardware [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 750.016166] env[61987]: DEBUG nova.virt.hardware [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 750.016166] env[61987]: DEBUG nova.virt.hardware [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 750.016166] env[61987]: DEBUG nova.virt.hardware [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 750.016642] env[61987]: DEBUG nova.virt.hardware [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 750.016642] env[61987]: DEBUG nova.virt.hardware [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 750.016642] env[61987]: DEBUG nova.virt.hardware [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 750.016990] env[61987]: DEBUG oslo_vmware.api [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 750.016990] env[61987]: value = "task-1061512" [ 750.016990] env[61987]: _type = "Task" [ 750.016990] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.017743] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4ae50d-82bc-4586-95d1-90808b03c9d6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.021670] env[61987]: DEBUG nova.scheduler.client.report [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 750.041369] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ccef6d5-ceeb-48eb-8232-8ff4affc9a8d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.044449] env[61987]: DEBUG oslo_vmware.api [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061512, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.155256] env[61987]: DEBUG nova.network.neutron [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 750.367199] env[61987]: DEBUG nova.network.neutron [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Updating instance_info_cache with network_info: [{"id": "55e46e63-9076-400e-a73a-19fc4bb21bec", "address": "fa:16:3e:09:61:0c", "network": {"id": "59d8d7d6-fd71-4dec-802e-e0f9e670ff70", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1845227031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "315a2a3450024ed38e27cd7758f392f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55e46e63-90", "ovs_interfaceid": "55e46e63-9076-400e-a73a-19fc4bb21bec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.526199] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.093s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.526420] env[61987]: INFO nova.compute.manager [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Migrating [ 750.526662] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.527582] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.531539] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.380s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.531810] env[61987]: DEBUG nova.objects.instance [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Lazy-loading 'resources' on Instance uuid e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 750.539864] env[61987]: DEBUG oslo_vmware.api [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061512, 'name': ReconfigVM_Task, 'duration_secs': 0.181268} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.540019] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234299', 'volume_id': 'fcdd03f8-ce63-4537-a4ea-aed9ed31ca09', 'name': 'volume-fcdd03f8-ce63-4537-a4ea-aed9ed31ca09', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '325183b9-e35b-4f9c-a271-d0fdb05288bb', 'attached_at': '', 'detached_at': '', 'volume_id': 'fcdd03f8-ce63-4537-a4ea-aed9ed31ca09', 'serial': 'fcdd03f8-ce63-4537-a4ea-aed9ed31ca09'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 750.540305] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 750.541522] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93ba9f52-7a36-4853-acda-4c9b85e01b22 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.549944] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 750.549944] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6d32d849-c11d-4ad5-8946-cf015e2c4138 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.645606] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 750.645606] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 750.645873] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Deleting the datastore file [datastore1] 325183b9-e35b-4f9c-a271-d0fdb05288bb {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 750.646059] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ea167894-be4b-4425-b6da-561f332ca282 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.653138] env[61987]: DEBUG oslo_vmware.api [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 750.653138] env[61987]: value = "task-1061514" [ 750.653138] env[61987]: _type = "Task" [ 750.653138] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.661779] env[61987]: DEBUG oslo_vmware.api [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061514, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.850760] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 750.853640] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5142339a-826e-467c-999e-b415af563dc6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.860193] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Waiting for the task: (returnval){ [ 750.860193] env[61987]: value = "task-1061515" [ 750.860193] env[61987]: _type = "Task" [ 750.860193] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.871269] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Releasing lock "refresh_cache-8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.871269] env[61987]: DEBUG nova.compute.manager [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Instance network_info: |[{"id": "55e46e63-9076-400e-a73a-19fc4bb21bec", "address": "fa:16:3e:09:61:0c", "network": {"id": "59d8d7d6-fd71-4dec-802e-e0f9e670ff70", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1845227031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "315a2a3450024ed38e27cd7758f392f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55e46e63-90", "ovs_interfaceid": "55e46e63-9076-400e-a73a-19fc4bb21bec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 750.871269] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': task-1061515, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.871269] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:61:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '028bae2d-fe6c-4207-b4a3-3fab45fbf1d6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '55e46e63-9076-400e-a73a-19fc4bb21bec', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 750.879652] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Creating folder: Project (315a2a3450024ed38e27cd7758f392f9). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 750.879652] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f29b1d1a-6489-43c4-bddc-f15854a4e0eb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.891261] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Created folder: Project (315a2a3450024ed38e27cd7758f392f9) in parent group-v234219. [ 750.891482] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Creating folder: Instances. Parent ref: group-v234303. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 750.891751] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-215f9e13-08a9-4ea4-8ff3-d7e7c1a20fd3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.903101] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Created folder: Instances in parent group-v234303. [ 750.903904] env[61987]: DEBUG oslo.service.loopingcall [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 750.903904] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 750.903904] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2c7655b9-d000-4135-a896-282980a477e0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.925779] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 750.925779] env[61987]: value = "task-1061518" [ 750.925779] env[61987]: _type = "Task" [ 750.925779] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.938168] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061518, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.952550] env[61987]: DEBUG nova.compute.manager [req-c2303642-9db0-41fa-b196-f15c2eded3c0 req-6a236384-9170-47b6-9db0-1ad84bc7e8a0 service nova] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Received event network-changed-55e46e63-9076-400e-a73a-19fc4bb21bec {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 750.952662] env[61987]: DEBUG nova.compute.manager [req-c2303642-9db0-41fa-b196-f15c2eded3c0 req-6a236384-9170-47b6-9db0-1ad84bc7e8a0 service nova] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Refreshing instance network info cache due to event network-changed-55e46e63-9076-400e-a73a-19fc4bb21bec. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 750.952871] env[61987]: DEBUG oslo_concurrency.lockutils [req-c2303642-9db0-41fa-b196-f15c2eded3c0 req-6a236384-9170-47b6-9db0-1ad84bc7e8a0 service nova] Acquiring lock "refresh_cache-8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.953199] env[61987]: DEBUG oslo_concurrency.lockutils [req-c2303642-9db0-41fa-b196-f15c2eded3c0 req-6a236384-9170-47b6-9db0-1ad84bc7e8a0 service nova] Acquired lock "refresh_cache-8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.953199] env[61987]: DEBUG nova.network.neutron [req-c2303642-9db0-41fa-b196-f15c2eded3c0 req-6a236384-9170-47b6-9db0-1ad84bc7e8a0 service nova] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Refreshing network info cache for port 55e46e63-9076-400e-a73a-19fc4bb21bec {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 751.033922] env[61987]: INFO nova.compute.rpcapi [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Automatically selected compute RPC version 6.4 from minimum service version 68 [ 751.034681] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.164899] env[61987]: DEBUG oslo_vmware.api [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061514, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199017} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.168610] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 751.168917] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 751.169543] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 751.169724] env[61987]: INFO nova.compute.manager [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Took 2.38 seconds to destroy the instance on the hypervisor. [ 751.169982] env[61987]: DEBUG oslo.service.loopingcall [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 751.170970] env[61987]: DEBUG nova.compute.manager [req-b4447dee-75ea-4c9a-80f0-37a3ef1786ae req-4519b9fe-7750-4d66-904c-792e6e1f5d3a service nova] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Received event network-vif-plugged-04dbba0f-63b4-4d42-a3ec-7fe8dbd70488 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 751.171185] env[61987]: DEBUG oslo_concurrency.lockutils [req-b4447dee-75ea-4c9a-80f0-37a3ef1786ae req-4519b9fe-7750-4d66-904c-792e6e1f5d3a service nova] Acquiring lock "bb7243dc-a950-4d3e-9b17-944e57cdd56f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.171400] env[61987]: DEBUG oslo_concurrency.lockutils [req-b4447dee-75ea-4c9a-80f0-37a3ef1786ae req-4519b9fe-7750-4d66-904c-792e6e1f5d3a service nova] Lock "bb7243dc-a950-4d3e-9b17-944e57cdd56f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.171601] env[61987]: DEBUG oslo_concurrency.lockutils [req-b4447dee-75ea-4c9a-80f0-37a3ef1786ae req-4519b9fe-7750-4d66-904c-792e6e1f5d3a service nova] Lock "bb7243dc-a950-4d3e-9b17-944e57cdd56f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.171749] env[61987]: DEBUG nova.compute.manager [req-b4447dee-75ea-4c9a-80f0-37a3ef1786ae req-4519b9fe-7750-4d66-904c-792e6e1f5d3a service nova] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] No waiting events found dispatching network-vif-plugged-04dbba0f-63b4-4d42-a3ec-7fe8dbd70488 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 751.171918] env[61987]: WARNING nova.compute.manager [req-b4447dee-75ea-4c9a-80f0-37a3ef1786ae req-4519b9fe-7750-4d66-904c-792e6e1f5d3a service nova] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Received unexpected event network-vif-plugged-04dbba0f-63b4-4d42-a3ec-7fe8dbd70488 for instance with vm_state building and task_state spawning. [ 751.172988] env[61987]: DEBUG nova.compute.manager [-] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 751.173169] env[61987]: DEBUG nova.network.neutron [-] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 751.273372] env[61987]: DEBUG nova.network.neutron [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Successfully updated port: 04dbba0f-63b4-4d42-a3ec-7fe8dbd70488 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 751.373615] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': task-1061515, 'name': PowerOffVM_Task, 'duration_secs': 0.186445} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.373960] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 751.374197] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 751.374989] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b01b132-f5e0-477d-ac4d-41038dd2fc29 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.383883] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 751.384189] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-df56d72e-657f-441b-ba22-53cd00c635c6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.420033] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 751.420314] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 751.420564] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Deleting the datastore file [datastore1] 1afcffda-0bd2-4cbe-8cfb-12a91bb50975 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 751.421219] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-59b49743-d720-4a88-a578-d84c36094331 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.433777] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Waiting for the task: (returnval){ [ 751.433777] env[61987]: value = "task-1061520" [ 751.433777] env[61987]: _type = "Task" [ 751.433777] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.444727] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061518, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.451458] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': task-1061520, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.465769] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e2f608-45e4-4e46-a1f5-461be8218dc0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.474850] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c6a5730-d5b5-4950-bd8f-ca19a9843baa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.510400] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43afee1a-652c-4553-8214-93636c5d9d85 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.519281] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b8dd6b2-a3c5-4d8b-8fbf-fe3a40769d54 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.533909] env[61987]: DEBUG nova.compute.provider_tree [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.556062] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquiring lock "refresh_cache-80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.556062] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquired lock "refresh_cache-80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.556062] env[61987]: DEBUG nova.network.neutron [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 751.703453] env[61987]: DEBUG nova.network.neutron [req-c2303642-9db0-41fa-b196-f15c2eded3c0 req-6a236384-9170-47b6-9db0-1ad84bc7e8a0 service nova] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Updated VIF entry in instance network info cache for port 55e46e63-9076-400e-a73a-19fc4bb21bec. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 751.703945] env[61987]: DEBUG nova.network.neutron [req-c2303642-9db0-41fa-b196-f15c2eded3c0 req-6a236384-9170-47b6-9db0-1ad84bc7e8a0 service nova] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Updating instance_info_cache with network_info: [{"id": "55e46e63-9076-400e-a73a-19fc4bb21bec", "address": "fa:16:3e:09:61:0c", "network": {"id": "59d8d7d6-fd71-4dec-802e-e0f9e670ff70", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1845227031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "315a2a3450024ed38e27cd7758f392f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55e46e63-90", "ovs_interfaceid": "55e46e63-9076-400e-a73a-19fc4bb21bec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.774665] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Acquiring lock "refresh_cache-bb7243dc-a950-4d3e-9b17-944e57cdd56f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.774870] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Acquired lock "refresh_cache-bb7243dc-a950-4d3e-9b17-944e57cdd56f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.775465] env[61987]: DEBUG nova.network.neutron [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 751.934013] env[61987]: DEBUG nova.network.neutron [-] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.940673] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061518, 'name': CreateVM_Task, 'duration_secs': 0.596068} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.945047] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 751.945988] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.946254] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.946992] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 751.947800] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-940e430f-8d8d-46fd-ac68-808910524352 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.953692] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': task-1061520, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.106224} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.953952] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 751.954171] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 751.954359] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 751.958124] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 751.958124] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]524bb4bc-d0ef-691a-bc04-006115ac35be" [ 751.958124] env[61987]: _type = "Task" [ 751.958124] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.966277] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524bb4bc-d0ef-691a-bc04-006115ac35be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.037408] env[61987]: DEBUG nova.scheduler.client.report [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 752.206103] env[61987]: DEBUG oslo_concurrency.lockutils [req-c2303642-9db0-41fa-b196-f15c2eded3c0 req-6a236384-9170-47b6-9db0-1ad84bc7e8a0 service nova] Releasing lock "refresh_cache-8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.206369] env[61987]: DEBUG nova.compute.manager [req-c2303642-9db0-41fa-b196-f15c2eded3c0 req-6a236384-9170-47b6-9db0-1ad84bc7e8a0 service nova] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Received event network-changed-021ce2e6-14d2-4f08-b654-6770827da9ee {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 752.206544] env[61987]: DEBUG nova.compute.manager [req-c2303642-9db0-41fa-b196-f15c2eded3c0 req-6a236384-9170-47b6-9db0-1ad84bc7e8a0 service nova] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Refreshing instance network info cache due to event network-changed-021ce2e6-14d2-4f08-b654-6770827da9ee. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 752.206770] env[61987]: DEBUG oslo_concurrency.lockutils [req-c2303642-9db0-41fa-b196-f15c2eded3c0 req-6a236384-9170-47b6-9db0-1ad84bc7e8a0 service nova] Acquiring lock "refresh_cache-95641d0b-970c-4b94-8568-5c46d0808345" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.206920] env[61987]: DEBUG oslo_concurrency.lockutils [req-c2303642-9db0-41fa-b196-f15c2eded3c0 req-6a236384-9170-47b6-9db0-1ad84bc7e8a0 service nova] Acquired lock "refresh_cache-95641d0b-970c-4b94-8568-5c46d0808345" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.207097] env[61987]: DEBUG nova.network.neutron [req-c2303642-9db0-41fa-b196-f15c2eded3c0 req-6a236384-9170-47b6-9db0-1ad84bc7e8a0 service nova] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Refreshing network info cache for port 021ce2e6-14d2-4f08-b654-6770827da9ee {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 752.287381] env[61987]: DEBUG nova.network.neutron [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Updating instance_info_cache with network_info: [{"id": "fc1b4b12-b396-40ac-9eab-a1309f2259e4", "address": "fa:16:3e:ec:35:ab", "network": {"id": "8a9770a4-fed3-4232-ac78-8a1fc2223229", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.118", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "705320a8b6a549d0b197ee1a35404d52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bafe8721-91d4-4127-b215-d9e8e27947dc", "external-id": "nsx-vlan-transportzone-680", "segmentation_id": 680, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc1b4b12-b3", "ovs_interfaceid": "fc1b4b12-b396-40ac-9eab-a1309f2259e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.324318] env[61987]: DEBUG nova.network.neutron [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 752.442066] env[61987]: INFO nova.compute.manager [-] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Took 1.27 seconds to deallocate network for instance. [ 752.471730] env[61987]: DEBUG nova.network.neutron [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Updating instance_info_cache with network_info: [{"id": "04dbba0f-63b4-4d42-a3ec-7fe8dbd70488", "address": "fa:16:3e:c0:8f:1a", "network": {"id": "d4f3153c-f5e1-44ec-9f6c-b8f6d181553f", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2075691453-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a53eb8ad8274fe18fd149c44eb0ffbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa09e855-8af1-419b-b78d-8ffcc94b1bfb", "external-id": "nsx-vlan-transportzone-901", "segmentation_id": 901, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04dbba0f-63", "ovs_interfaceid": "04dbba0f-63b4-4d42-a3ec-7fe8dbd70488", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.485780] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524bb4bc-d0ef-691a-bc04-006115ac35be, 'name': SearchDatastore_Task, 'duration_secs': 0.009748} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.486103] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.486426] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 752.486685] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.487292] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.487562] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 752.487863] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a8f7a1eb-ef0b-4a38-9d26-f1136f720ca8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.498156] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 752.498360] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 752.499112] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07e316a5-7a4a-465d-af24-879389505364 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.505250] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 752.505250] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52b30563-b777-9012-4af4-fd011f401b6d" [ 752.505250] env[61987]: _type = "Task" [ 752.505250] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.514139] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b30563-b777-9012-4af4-fd011f401b6d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.542037] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.010s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.545231] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.152s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.547131] env[61987]: INFO nova.compute.claims [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 752.568459] env[61987]: INFO nova.scheduler.client.report [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Deleted allocations for instance e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1 [ 752.789155] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Releasing lock "refresh_cache-80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.977934] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Releasing lock "refresh_cache-bb7243dc-a950-4d3e-9b17-944e57cdd56f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.978291] env[61987]: DEBUG nova.compute.manager [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Instance network_info: |[{"id": "04dbba0f-63b4-4d42-a3ec-7fe8dbd70488", "address": "fa:16:3e:c0:8f:1a", "network": {"id": "d4f3153c-f5e1-44ec-9f6c-b8f6d181553f", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2075691453-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a53eb8ad8274fe18fd149c44eb0ffbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa09e855-8af1-419b-b78d-8ffcc94b1bfb", "external-id": "nsx-vlan-transportzone-901", "segmentation_id": 901, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04dbba0f-63", "ovs_interfaceid": "04dbba0f-63b4-4d42-a3ec-7fe8dbd70488", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 752.981477] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:8f:1a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aa09e855-8af1-419b-b78d-8ffcc94b1bfb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '04dbba0f-63b4-4d42-a3ec-7fe8dbd70488', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 752.989577] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Creating folder: Project (5a53eb8ad8274fe18fd149c44eb0ffbb). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 752.994571] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-21a83e32-c168-480c-9839-995ab5bbf0ae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.997413] env[61987]: DEBUG nova.compute.manager [req-5dc694bd-a1de-46a6-8272-6ba7a8b001c2 req-e4724a2a-be1d-4ee8-9010-3f4e93864032 service nova] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Received event network-vif-deleted-853aa7a2-96d2-4665-b680-811dffc26bfe {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 753.004636] env[61987]: DEBUG nova.virt.hardware [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 753.004887] env[61987]: DEBUG nova.virt.hardware [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 753.005548] env[61987]: DEBUG nova.virt.hardware [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 753.005548] env[61987]: DEBUG nova.virt.hardware [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 753.005548] env[61987]: DEBUG nova.virt.hardware [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 753.005548] env[61987]: DEBUG nova.virt.hardware [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 753.005831] env[61987]: DEBUG nova.virt.hardware [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 753.005962] env[61987]: DEBUG nova.virt.hardware [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 753.006106] env[61987]: DEBUG nova.virt.hardware [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 753.006296] env[61987]: DEBUG nova.virt.hardware [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 753.006478] env[61987]: DEBUG nova.virt.hardware [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 753.007699] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e2ebb9-1829-4700-8656-5b566ad1bb98 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.014563] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Created folder: Project (5a53eb8ad8274fe18fd149c44eb0ffbb) in parent group-v234219. [ 753.014786] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Creating folder: Instances. Parent ref: group-v234306. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 753.015721] env[61987]: INFO nova.compute.manager [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Took 0.57 seconds to detach 1 volumes for instance. [ 753.017653] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c511d932-1928-4970-acea-2f2d7087ab63 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.029758] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea5bb613-3d2b-4122-b0fe-e71d48eee570 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.033853] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b30563-b777-9012-4af4-fd011f401b6d, 'name': SearchDatastore_Task, 'duration_secs': 0.009857} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.035717] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Created folder: Instances in parent group-v234306. [ 753.035966] env[61987]: DEBUG oslo.service.loopingcall [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 753.036842] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3193fb46-238a-418f-b4c1-6da6500fba27 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.040364] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 753.050027] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c83e0dbd-cdf1-4a09-a5e4-bea747b6105f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.061178] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Instance VIF info [] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 753.066705] env[61987]: DEBUG oslo.service.loopingcall [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 753.067615] env[61987]: DEBUG nova.network.neutron [req-c2303642-9db0-41fa-b196-f15c2eded3c0 req-6a236384-9170-47b6-9db0-1ad84bc7e8a0 service nova] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Updated VIF entry in instance network info cache for port 021ce2e6-14d2-4f08-b654-6770827da9ee. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 753.067943] env[61987]: DEBUG nova.network.neutron [req-c2303642-9db0-41fa-b196-f15c2eded3c0 req-6a236384-9170-47b6-9db0-1ad84bc7e8a0 service nova] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Updating instance_info_cache with network_info: [{"id": "021ce2e6-14d2-4f08-b654-6770827da9ee", "address": "fa:16:3e:96:d4:75", "network": {"id": "801ddda8-4f5d-42fe-9660-5b294986ecce", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-764406404-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.199", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e6fb7bd076b41838cf2a873b3e4b8a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe20ef0e-0991-44d7-887d-08dddac0b56b", "external-id": "nsx-vlan-transportzone-991", "segmentation_id": 991, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap021ce2e6-14", "ovs_interfaceid": "021ce2e6-14d2-4f08-b654-6770827da9ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.072743] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 753.073109] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ec200003-c7a3-48b4-a10c-14ba4be0a3d0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.091885] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 753.091885] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5251ff73-941f-2297-e4a6-9fd9d63f954f" [ 753.091885] env[61987]: _type = "Task" [ 753.091885] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.092196] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 753.092196] env[61987]: value = "task-1061523" [ 753.092196] env[61987]: _type = "Task" [ 753.092196] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.092524] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3a0444b2-b23e-4046-b4cc-ade2fb219ae9 tempest-ServerAddressesTestJSON-1923041393 tempest-ServerAddressesTestJSON-1923041393-project-member] Lock "e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.628s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.101891] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 753.101891] env[61987]: value = "task-1061524" [ 753.101891] env[61987]: _type = "Task" [ 753.101891] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.118895] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5251ff73-941f-2297-e4a6-9fd9d63f954f, 'name': SearchDatastore_Task, 'duration_secs': 0.009834} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.119438] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061523, 'name': CreateVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.119633] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061524, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.119867] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.120164] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2/8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 753.120428] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9f968f93-039a-4e96-9156-17da753ae0ca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.128077] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 753.128077] env[61987]: value = "task-1061525" [ 753.128077] env[61987]: _type = "Task" [ 753.128077] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.136489] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061525, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.223394] env[61987]: DEBUG nova.compute.manager [req-6d6dd1e5-724c-403b-86c7-13a2cd47ce4c req-39eaae57-691d-41e3-91a1-d07c60d762c3 service nova] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Received event network-changed-04dbba0f-63b4-4d42-a3ec-7fe8dbd70488 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 753.223623] env[61987]: DEBUG nova.compute.manager [req-6d6dd1e5-724c-403b-86c7-13a2cd47ce4c req-39eaae57-691d-41e3-91a1-d07c60d762c3 service nova] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Refreshing instance network info cache due to event network-changed-04dbba0f-63b4-4d42-a3ec-7fe8dbd70488. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 753.223879] env[61987]: DEBUG oslo_concurrency.lockutils [req-6d6dd1e5-724c-403b-86c7-13a2cd47ce4c req-39eaae57-691d-41e3-91a1-d07c60d762c3 service nova] Acquiring lock "refresh_cache-bb7243dc-a950-4d3e-9b17-944e57cdd56f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.224055] env[61987]: DEBUG oslo_concurrency.lockutils [req-6d6dd1e5-724c-403b-86c7-13a2cd47ce4c req-39eaae57-691d-41e3-91a1-d07c60d762c3 service nova] Acquired lock "refresh_cache-bb7243dc-a950-4d3e-9b17-944e57cdd56f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.224230] env[61987]: DEBUG nova.network.neutron [req-6d6dd1e5-724c-403b-86c7-13a2cd47ce4c req-39eaae57-691d-41e3-91a1-d07c60d762c3 service nova] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Refreshing network info cache for port 04dbba0f-63b4-4d42-a3ec-7fe8dbd70488 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 753.525235] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.576965] env[61987]: DEBUG oslo_concurrency.lockutils [req-c2303642-9db0-41fa-b196-f15c2eded3c0 req-6a236384-9170-47b6-9db0-1ad84bc7e8a0 service nova] Releasing lock "refresh_cache-95641d0b-970c-4b94-8568-5c46d0808345" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.611205] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061523, 'name': CreateVM_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.619703] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061524, 'name': CreateVM_Task, 'duration_secs': 0.351976} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.619899] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 753.620355] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.620529] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.620876] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 753.621170] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45c678f5-e58e-4318-b524-9d0f268a22fe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.630128] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Waiting for the task: (returnval){ [ 753.630128] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52697231-976d-3c11-e566-0cbf2dd995e4" [ 753.630128] env[61987]: _type = "Task" [ 753.630128] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.648015] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061525, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.491534} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.648291] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52697231-976d-3c11-e566-0cbf2dd995e4, 'name': SearchDatastore_Task, 'duration_secs': 0.013248} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.648528] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2/8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 753.648743] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 753.649068] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.649288] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 753.649519] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.649667] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.649846] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 753.650118] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b97bc06c-d18e-4537-a07a-f69d50bab4e5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.652028] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-afdf9e6f-8c50-406e-8ecb-62a8f1d6a59d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.666026] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 753.666026] env[61987]: value = "task-1061526" [ 753.666026] env[61987]: _type = "Task" [ 753.666026] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.666026] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 753.666026] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 753.671722] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73e1ec6d-c258-4eeb-8fcf-467713bed3a8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.681068] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Waiting for the task: (returnval){ [ 753.681068] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52ddb0fd-488d-35c8-3ea0-73b19a472e20" [ 753.681068] env[61987]: _type = "Task" [ 753.681068] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.682895] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061526, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.694845] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52ddb0fd-488d-35c8-3ea0-73b19a472e20, 'name': SearchDatastore_Task, 'duration_secs': 0.010881} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.695867] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2cc68e41-7110-4209-9ca0-9bfd9dd91874 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.704971] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Waiting for the task: (returnval){ [ 753.704971] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52cff7aa-81e6-9b0f-acec-b037c2f703be" [ 753.704971] env[61987]: _type = "Task" [ 753.704971] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.716802] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52cff7aa-81e6-9b0f-acec-b037c2f703be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.991398] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f9dd294-dbad-4421-afb8-bef4d986da34 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.003433] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d81ed63e-17a6-4ea1-ae13-d6af03d77229 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.035088] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddae7c02-24b8-4309-99bb-6dcc1b4bb066 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.043789] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d7c48b-18e7-440a-82f0-1a8d0abc7ff6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.057769] env[61987]: DEBUG nova.compute.provider_tree [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.109382] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061523, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.140056] env[61987]: DEBUG nova.network.neutron [req-6d6dd1e5-724c-403b-86c7-13a2cd47ce4c req-39eaae57-691d-41e3-91a1-d07c60d762c3 service nova] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Updated VIF entry in instance network info cache for port 04dbba0f-63b4-4d42-a3ec-7fe8dbd70488. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 754.140441] env[61987]: DEBUG nova.network.neutron [req-6d6dd1e5-724c-403b-86c7-13a2cd47ce4c req-39eaae57-691d-41e3-91a1-d07c60d762c3 service nova] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Updating instance_info_cache with network_info: [{"id": "04dbba0f-63b4-4d42-a3ec-7fe8dbd70488", "address": "fa:16:3e:c0:8f:1a", "network": {"id": "d4f3153c-f5e1-44ec-9f6c-b8f6d181553f", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-2075691453-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a53eb8ad8274fe18fd149c44eb0ffbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa09e855-8af1-419b-b78d-8ffcc94b1bfb", "external-id": "nsx-vlan-transportzone-901", "segmentation_id": 901, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04dbba0f-63", "ovs_interfaceid": "04dbba0f-63b4-4d42-a3ec-7fe8dbd70488", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.174896] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061526, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073806} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.175201] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 754.176159] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beb77bd3-7068-4289-91da-716480485016 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.199265] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2/8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 754.199564] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-52f79e29-92cf-4e7d-a823-8bfd7f2944d8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.230295] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 754.230295] env[61987]: value = "task-1061527" [ 754.230295] env[61987]: _type = "Task" [ 754.230295] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.230295] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52cff7aa-81e6-9b0f-acec-b037c2f703be, 'name': SearchDatastore_Task, 'duration_secs': 0.010567} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.230295] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.230570] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 1afcffda-0bd2-4cbe-8cfb-12a91bb50975/1afcffda-0bd2-4cbe-8cfb-12a91bb50975.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 754.234512] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dfd13709-dcd1-46fb-a208-6286c498869c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.243887] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061527, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.245459] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Waiting for the task: (returnval){ [ 754.245459] env[61987]: value = "task-1061528" [ 754.245459] env[61987]: _type = "Task" [ 754.245459] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.255702] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': task-1061528, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.305808] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-680df39c-195b-4abf-922f-239b1ecc8f66 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.325757] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Updating instance '80f13fca-2df1-4aa0-96f1-b61a60c4fdc6' progress to 0 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 754.562022] env[61987]: DEBUG nova.scheduler.client.report [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 754.605933] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061523, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.643428] env[61987]: DEBUG oslo_concurrency.lockutils [req-6d6dd1e5-724c-403b-86c7-13a2cd47ce4c req-39eaae57-691d-41e3-91a1-d07c60d762c3 service nova] Releasing lock "refresh_cache-bb7243dc-a950-4d3e-9b17-944e57cdd56f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.742232] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061527, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.755924] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': task-1061528, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.832585] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 754.832969] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3a3f2ec9-376e-46aa-b484-81fdc19a88e1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.841794] env[61987]: DEBUG oslo_vmware.api [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for the task: (returnval){ [ 754.841794] env[61987]: value = "task-1061529" [ 754.841794] env[61987]: _type = "Task" [ 754.841794] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.851935] env[61987]: DEBUG oslo_vmware.api [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061529, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.070538] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.526s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.070965] env[61987]: DEBUG nova.compute.manager [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 755.076327] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.726s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.076327] env[61987]: INFO nova.compute.claims [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 755.116785] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061523, 'name': CreateVM_Task, 'duration_secs': 1.676135} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.116785] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 755.117460] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.117828] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.118171] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 755.118518] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37e2d577-21c6-4e91-aa29-662b106e4ac1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.125671] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Waiting for the task: (returnval){ [ 755.125671] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]522074fc-b4ce-b8e5-4101-109a72653bd2" [ 755.125671] env[61987]: _type = "Task" [ 755.125671] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.140282] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]522074fc-b4ce-b8e5-4101-109a72653bd2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.243058] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061527, 'name': ReconfigVM_Task, 'duration_secs': 0.674169} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.243356] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Reconfigured VM instance instance-00000034 to attach disk [datastore2] 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2/8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 755.246631] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57307c46-75f1-4ce5-be25-1f821ded100f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.252039] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 755.252039] env[61987]: value = "task-1061530" [ 755.252039] env[61987]: _type = "Task" [ 755.252039] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.258836] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': task-1061528, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.862618} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.259534] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 1afcffda-0bd2-4cbe-8cfb-12a91bb50975/1afcffda-0bd2-4cbe-8cfb-12a91bb50975.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 755.259789] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 755.260073] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ca68f766-5495-4a46-8118-4d6484e417e7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.265240] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061530, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.270852] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Waiting for the task: (returnval){ [ 755.270852] env[61987]: value = "task-1061531" [ 755.270852] env[61987]: _type = "Task" [ 755.270852] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.279380] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': task-1061531, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.354010] env[61987]: DEBUG oslo_vmware.api [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061529, 'name': PowerOffVM_Task, 'duration_secs': 0.27478} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.354010] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 755.354010] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Updating instance '80f13fca-2df1-4aa0-96f1-b61a60c4fdc6' progress to 17 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 755.583325] env[61987]: DEBUG nova.compute.utils [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 755.587801] env[61987]: DEBUG nova.compute.manager [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 755.588238] env[61987]: DEBUG nova.network.neutron [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 755.637735] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]522074fc-b4ce-b8e5-4101-109a72653bd2, 'name': SearchDatastore_Task, 'duration_secs': 0.016594} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.638079] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.638341] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 755.638577] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.638734] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.638919] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 755.639200] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0129eefb-ebb4-4132-a0b4-ae9ba4d3306c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.660605] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 755.660784] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 755.661572] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-278d1d6d-b34d-4aef-a457-9961f3fb0547 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.668629] env[61987]: DEBUG nova.policy [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0945d6196b8c473da5308e6d784ab479', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b77fbbfc7f6437083f9f30e8fdeeb35', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 755.671670] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Waiting for the task: (returnval){ [ 755.671670] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5230bba4-0bbd-a92b-5ff9-9b7387f20775" [ 755.671670] env[61987]: _type = "Task" [ 755.671670] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.680771] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5230bba4-0bbd-a92b-5ff9-9b7387f20775, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.764419] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061530, 'name': Rename_Task, 'duration_secs': 0.154713} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.765420] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 755.766241] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-98180db0-e2bf-44aa-8786-becf650dda7a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.778580] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 755.778580] env[61987]: value = "task-1061532" [ 755.778580] env[61987]: _type = "Task" [ 755.778580] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.788627] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': task-1061531, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066265} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.790673] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 755.797321] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcbc7c87-35cd-4dcc-8443-7a297a6a3965 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.798933] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061532, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.821981] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] 1afcffda-0bd2-4cbe-8cfb-12a91bb50975/1afcffda-0bd2-4cbe-8cfb-12a91bb50975.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 755.821981] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c136baf-1d11-4ed6-8802-f14265bb6692 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.845835] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Waiting for the task: (returnval){ [ 755.845835] env[61987]: value = "task-1061533" [ 755.845835] env[61987]: _type = "Task" [ 755.845835] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.860666] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 755.860969] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 755.861174] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 755.861388] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 755.861581] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 755.861853] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 755.861982] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 755.862168] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 755.862345] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 755.862547] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 755.862784] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 755.868192] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': task-1061533, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.868884] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-704dd80e-fded-455f-ae17-01fbd87f94c6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.888388] env[61987]: DEBUG oslo_vmware.api [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for the task: (returnval){ [ 755.888388] env[61987]: value = "task-1061534" [ 755.888388] env[61987]: _type = "Task" [ 755.888388] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.898156] env[61987]: DEBUG oslo_vmware.api [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061534, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.091481] env[61987]: DEBUG nova.compute.manager [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 756.098439] env[61987]: DEBUG nova.network.neutron [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Successfully created port: ac7175cc-c93a-4a76-aa42-faf9d1c022a1 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 756.188780] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5230bba4-0bbd-a92b-5ff9-9b7387f20775, 'name': SearchDatastore_Task, 'duration_secs': 0.010812} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.189643] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42f327f5-f62c-47a6-b014-61878018b483 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.202730] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Waiting for the task: (returnval){ [ 756.202730] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]523082a4-fb58-ae9b-66b7-3d47aa911c45" [ 756.202730] env[61987]: _type = "Task" [ 756.202730] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.213557] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523082a4-fb58-ae9b-66b7-3d47aa911c45, 'name': SearchDatastore_Task} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.213856] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.214139] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] bb7243dc-a950-4d3e-9b17-944e57cdd56f/bb7243dc-a950-4d3e-9b17-944e57cdd56f.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 756.214473] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a7a3aa12-7ea8-418b-807a-d70a5df9a82f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.222570] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Waiting for the task: (returnval){ [ 756.222570] env[61987]: value = "task-1061535" [ 756.222570] env[61987]: _type = "Task" [ 756.222570] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.232820] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': task-1061535, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.297173] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061532, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.360180] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': task-1061533, 'name': ReconfigVM_Task, 'duration_secs': 0.293894} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.364412] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Reconfigured VM instance instance-00000032 to attach disk [datastore1] 1afcffda-0bd2-4cbe-8cfb-12a91bb50975/1afcffda-0bd2-4cbe-8cfb-12a91bb50975.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 756.365326] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7bc5fb2d-a497-4c79-9754-50e65c9c9786 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.373809] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Waiting for the task: (returnval){ [ 756.373809] env[61987]: value = "task-1061536" [ 756.373809] env[61987]: _type = "Task" [ 756.373809] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.387785] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': task-1061536, 'name': Rename_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.398502] env[61987]: DEBUG oslo_vmware.api [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061534, 'name': ReconfigVM_Task, 'duration_secs': 0.295636} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.401738] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Updating instance '80f13fca-2df1-4aa0-96f1-b61a60c4fdc6' progress to 33 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 756.585478] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4d2cd1-a028-437b-8e9b-887a6fb18fcf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.599721] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad96f68d-924a-4912-8c44-84073bf85dd5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.648159] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad9986f4-b036-4028-aeb2-cc78396ff5f6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.658663] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64dbc79e-e579-4ac7-8333-c59be06c6cfa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.676600] env[61987]: DEBUG nova.compute.provider_tree [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.734309] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': task-1061535, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500239} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.734665] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] bb7243dc-a950-4d3e-9b17-944e57cdd56f/bb7243dc-a950-4d3e-9b17-944e57cdd56f.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 756.734907] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 756.735189] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-222d01ff-0f78-41cd-80de-4e9fce1315e3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.742287] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Waiting for the task: (returnval){ [ 756.742287] env[61987]: value = "task-1061537" [ 756.742287] env[61987]: _type = "Task" [ 756.742287] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.751101] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': task-1061537, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.796103] env[61987]: DEBUG oslo_vmware.api [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061532, 'name': PowerOnVM_Task, 'duration_secs': 0.806994} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.797739] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 756.798048] env[61987]: INFO nova.compute.manager [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Took 8.90 seconds to spawn the instance on the hypervisor. [ 756.798824] env[61987]: DEBUG nova.compute.manager [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 756.799975] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e5c197-c21f-4c12-9418-cf37ff673c1d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.886429] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': task-1061536, 'name': Rename_Task, 'duration_secs': 0.245501} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.886429] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 756.886429] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-786384c7-1017-4fe1-a243-51754f029284 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.894301] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Waiting for the task: (returnval){ [ 756.894301] env[61987]: value = "task-1061538" [ 756.894301] env[61987]: _type = "Task" [ 756.894301] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.910026] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:08:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='3ca61d0c-b87e-4949-982d-be5132df125d',id=36,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-2078042112',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 756.910324] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 756.910491] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 756.910680] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 756.910832] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 756.910987] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 756.911220] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 756.911418] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 756.911599] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 756.911925] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 756.912026] env[61987]: DEBUG nova.virt.hardware [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 756.918458] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Reconfiguring VM instance instance-0000002b to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 756.918801] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': task-1061538, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.919377] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e798eea-7e45-4fa0-af4a-7a84b265dd51 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.940578] env[61987]: DEBUG oslo_vmware.api [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for the task: (returnval){ [ 756.940578] env[61987]: value = "task-1061539" [ 756.940578] env[61987]: _type = "Task" [ 756.940578] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.950152] env[61987]: DEBUG oslo_vmware.api [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061539, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.115975] env[61987]: DEBUG nova.compute.manager [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 757.142499] env[61987]: DEBUG nova.virt.hardware [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 757.142773] env[61987]: DEBUG nova.virt.hardware [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 757.142943] env[61987]: DEBUG nova.virt.hardware [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 757.143155] env[61987]: DEBUG nova.virt.hardware [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 757.143313] env[61987]: DEBUG nova.virt.hardware [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 757.143467] env[61987]: DEBUG nova.virt.hardware [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 757.143678] env[61987]: DEBUG nova.virt.hardware [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 757.143882] env[61987]: DEBUG nova.virt.hardware [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 757.146179] env[61987]: DEBUG nova.virt.hardware [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 757.146434] env[61987]: DEBUG nova.virt.hardware [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 757.146633] env[61987]: DEBUG nova.virt.hardware [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 757.147516] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c3ad798-7ca0-40b0-9878-181c01d29290 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.156098] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c09afe3a-9044-4802-8333-47faa1c6423f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.179697] env[61987]: DEBUG nova.scheduler.client.report [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 757.253101] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': task-1061537, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072518} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.253315] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 757.254120] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe6d247d-b88e-4c6d-9aa5-1e6f02750ac4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.276939] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] bb7243dc-a950-4d3e-9b17-944e57cdd56f/bb7243dc-a950-4d3e-9b17-944e57cdd56f.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 757.277130] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db5d8be6-2ac4-42f3-8180-0e1355df6ca0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.298138] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Waiting for the task: (returnval){ [ 757.298138] env[61987]: value = "task-1061540" [ 757.298138] env[61987]: _type = "Task" [ 757.298138] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.306820] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': task-1061540, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.322719] env[61987]: INFO nova.compute.manager [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Took 33.03 seconds to build instance. [ 757.405337] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': task-1061538, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.453082] env[61987]: DEBUG oslo_vmware.api [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061539, 'name': ReconfigVM_Task, 'duration_secs': 0.198366} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.453328] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Reconfigured VM instance instance-0000002b to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 757.454148] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f46723-cc5d-4c08-a5e8-8ee531e7977a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.479289] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Reconfiguring VM instance instance-0000002b to attach disk [datastore1] 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6/80f13fca-2df1-4aa0-96f1-b61a60c4fdc6.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 757.479653] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e369ad5-bdbc-4ee7-8065-2ef581cc252e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.501081] env[61987]: DEBUG oslo_vmware.api [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for the task: (returnval){ [ 757.501081] env[61987]: value = "task-1061541" [ 757.501081] env[61987]: _type = "Task" [ 757.501081] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.510119] env[61987]: DEBUG oslo_vmware.api [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061541, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.685193] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.611s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.685735] env[61987]: DEBUG nova.compute.manager [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 757.688444] env[61987]: DEBUG oslo_concurrency.lockutils [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.324s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.688676] env[61987]: DEBUG nova.objects.instance [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lazy-loading 'resources' on Instance uuid 02fe490b-75c8-406b-98e0-aa200f249673 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 757.809814] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': task-1061540, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.824632] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6472088-99bb-43e2-acf5-1490c41724c8 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.258s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.907175] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': task-1061538, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.012508] env[61987]: DEBUG oslo_vmware.api [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061541, 'name': ReconfigVM_Task, 'duration_secs': 0.287364} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.013394] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Reconfigured VM instance instance-0000002b to attach disk [datastore1] 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6/80f13fca-2df1-4aa0-96f1-b61a60c4fdc6.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 758.013394] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Updating instance '80f13fca-2df1-4aa0-96f1-b61a60c4fdc6' progress to 50 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 758.060349] env[61987]: DEBUG nova.compute.manager [req-bd546d39-b9f6-46d1-9acd-f6922b0a2e9d req-0f0a6ecf-6efb-4535-bb64-81c1cea33f76 service nova] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Received event network-vif-plugged-ac7175cc-c93a-4a76-aa42-faf9d1c022a1 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 758.060575] env[61987]: DEBUG oslo_concurrency.lockutils [req-bd546d39-b9f6-46d1-9acd-f6922b0a2e9d req-0f0a6ecf-6efb-4535-bb64-81c1cea33f76 service nova] Acquiring lock "f937a5ec-a1d7-41d8-b998-fa18f545b304-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.060791] env[61987]: DEBUG oslo_concurrency.lockutils [req-bd546d39-b9f6-46d1-9acd-f6922b0a2e9d req-0f0a6ecf-6efb-4535-bb64-81c1cea33f76 service nova] Lock "f937a5ec-a1d7-41d8-b998-fa18f545b304-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.061014] env[61987]: DEBUG oslo_concurrency.lockutils [req-bd546d39-b9f6-46d1-9acd-f6922b0a2e9d req-0f0a6ecf-6efb-4535-bb64-81c1cea33f76 service nova] Lock "f937a5ec-a1d7-41d8-b998-fa18f545b304-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.061417] env[61987]: DEBUG nova.compute.manager [req-bd546d39-b9f6-46d1-9acd-f6922b0a2e9d req-0f0a6ecf-6efb-4535-bb64-81c1cea33f76 service nova] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] No waiting events found dispatching network-vif-plugged-ac7175cc-c93a-4a76-aa42-faf9d1c022a1 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 758.061598] env[61987]: WARNING nova.compute.manager [req-bd546d39-b9f6-46d1-9acd-f6922b0a2e9d req-0f0a6ecf-6efb-4535-bb64-81c1cea33f76 service nova] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Received unexpected event network-vif-plugged-ac7175cc-c93a-4a76-aa42-faf9d1c022a1 for instance with vm_state building and task_state spawning. [ 758.141684] env[61987]: DEBUG nova.network.neutron [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Successfully updated port: ac7175cc-c93a-4a76-aa42-faf9d1c022a1 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 758.192190] env[61987]: DEBUG nova.compute.utils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 758.199363] env[61987]: DEBUG nova.compute.manager [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 758.199600] env[61987]: DEBUG nova.network.neutron [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 758.254266] env[61987]: DEBUG nova.policy [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f9b90037d9384f549640bfcf344f0669', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '69749b752d1d4ad087f70f3d2ebe7b10', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 758.310527] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': task-1061540, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.328401] env[61987]: DEBUG nova.compute.manager [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 758.417040] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': task-1061538, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.522226] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30dac759-0fe4-4768-af21-2a96f6fbbb43 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.562693] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e3cf1d7-bf4f-4257-8744-dc77658a74b9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.594933] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Updating instance '80f13fca-2df1-4aa0-96f1-b61a60c4fdc6' progress to 67 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 758.646139] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "refresh_cache-f937a5ec-a1d7-41d8-b998-fa18f545b304" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.646480] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquired lock "refresh_cache-f937a5ec-a1d7-41d8-b998-fa18f545b304" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.646764] env[61987]: DEBUG nova.network.neutron [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 758.700120] env[61987]: DEBUG nova.compute.manager [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 758.721225] env[61987]: DEBUG nova.network.neutron [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Successfully created port: 8a3b1e33-f1b4-48d4-b5f3-bad9dedaeb70 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 758.796392] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d427279-0ae5-4fd7-b266-67e06bfc0c2b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.814558] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': task-1061540, 'name': ReconfigVM_Task, 'duration_secs': 1.263765} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.816667] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Reconfigured VM instance instance-00000035 to attach disk [datastore2] bb7243dc-a950-4d3e-9b17-944e57cdd56f/bb7243dc-a950-4d3e-9b17-944e57cdd56f.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 758.817361] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1fafd84d-98af-48a8-9afe-608ceb7b8c17 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.819759] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe21afcb-5229-4cfb-a3b0-a1e1fb30a72e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.858119] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c6be6fe-f20d-4f10-8743-9628b4b3e5c8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.861014] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Waiting for the task: (returnval){ [ 758.861014] env[61987]: value = "task-1061542" [ 758.861014] env[61987]: _type = "Task" [ 758.861014] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.871721] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcf7ff70-d49a-4d82-8082-72f9c12b2c9e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.878825] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': task-1061542, 'name': Rename_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.880968] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.895087] env[61987]: DEBUG nova.compute.provider_tree [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.907928] env[61987]: DEBUG oslo_vmware.api [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Task: {'id': task-1061538, 'name': PowerOnVM_Task, 'duration_secs': 1.839833} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.908944] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 758.908944] env[61987]: DEBUG nova.compute.manager [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 758.912376] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b4100d7-cb29-4445-a6bb-17c613181c12 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.181793] env[61987]: DEBUG nova.network.neutron [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Port fc1b4b12-b396-40ac-9eab-a1309f2259e4 binding to destination host cpu-1 is already ACTIVE {{(pid=61987) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 759.210823] env[61987]: DEBUG nova.network.neutron [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 759.375266] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': task-1061542, 'name': Rename_Task, 'duration_secs': 0.178324} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.375493] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 759.375745] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-36472ade-e993-42f1-9a95-1e32e9999503 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.384765] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Waiting for the task: (returnval){ [ 759.384765] env[61987]: value = "task-1061543" [ 759.384765] env[61987]: _type = "Task" [ 759.384765] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.395674] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': task-1061543, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.400323] env[61987]: DEBUG nova.scheduler.client.report [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 759.431659] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.449241] env[61987]: DEBUG nova.network.neutron [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Updating instance_info_cache with network_info: [{"id": "ac7175cc-c93a-4a76-aa42-faf9d1c022a1", "address": "fa:16:3e:4e:f6:f5", "network": {"id": "26d453d1-8edf-45e1-9e49-2a3225de6633", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1400564876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b77fbbfc7f6437083f9f30e8fdeeb35", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac7175cc-c9", "ovs_interfaceid": "ac7175cc-c93a-4a76-aa42-faf9d1c022a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.723957] env[61987]: DEBUG nova.compute.manager [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 759.753432] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 759.753736] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 759.754081] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 759.754164] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 759.754306] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 759.755098] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 759.755098] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 759.755098] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 759.755098] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 759.755259] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 759.755396] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 759.756523] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bce5fcd6-4492-4bdf-8eed-3fa0307ac2ab {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.767387] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fd484c0-88f5-4dfd-a4d5-d0171071f63a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.895942] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': task-1061543, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.906759] env[61987]: DEBUG oslo_concurrency.lockutils [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.217s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.908953] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.878s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.912113] env[61987]: INFO nova.compute.claims [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 759.932403] env[61987]: INFO nova.scheduler.client.report [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Deleted allocations for instance 02fe490b-75c8-406b-98e0-aa200f249673 [ 759.952469] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Releasing lock "refresh_cache-f937a5ec-a1d7-41d8-b998-fa18f545b304" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.952573] env[61987]: DEBUG nova.compute.manager [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Instance network_info: |[{"id": "ac7175cc-c93a-4a76-aa42-faf9d1c022a1", "address": "fa:16:3e:4e:f6:f5", "network": {"id": "26d453d1-8edf-45e1-9e49-2a3225de6633", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1400564876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b77fbbfc7f6437083f9f30e8fdeeb35", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac7175cc-c9", "ovs_interfaceid": "ac7175cc-c93a-4a76-aa42-faf9d1c022a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 759.953042] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4e:f6:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3734b156-0f7d-4721-b23c-d000412ec2eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ac7175cc-c93a-4a76-aa42-faf9d1c022a1', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 759.962397] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Creating folder: Project (7b77fbbfc7f6437083f9f30e8fdeeb35). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 759.965386] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-56836e56-b759-4752-9517-42b4c6d447ef {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.981982] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Created folder: Project (7b77fbbfc7f6437083f9f30e8fdeeb35) in parent group-v234219. [ 759.982207] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Creating folder: Instances. Parent ref: group-v234310. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 759.982516] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7ba87bfa-8555-4b2e-935e-443c51d72f13 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.996028] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Created folder: Instances in parent group-v234310. [ 759.996373] env[61987]: DEBUG oslo.service.loopingcall [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 759.998283] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 759.998283] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ec1130a7-e34f-475a-9ba0-dc54d0f58f41 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.022427] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 760.022427] env[61987]: value = "task-1061546" [ 760.022427] env[61987]: _type = "Task" [ 760.022427] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.034195] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061546, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.127369] env[61987]: DEBUG nova.compute.manager [req-b04dc650-830e-47fa-8f54-96000c4272e5 req-b30ea2ff-6a21-4df2-919b-a6a4ccb4dcf6 service nova] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Received event network-changed-ac7175cc-c93a-4a76-aa42-faf9d1c022a1 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 760.127580] env[61987]: DEBUG nova.compute.manager [req-b04dc650-830e-47fa-8f54-96000c4272e5 req-b30ea2ff-6a21-4df2-919b-a6a4ccb4dcf6 service nova] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Refreshing instance network info cache due to event network-changed-ac7175cc-c93a-4a76-aa42-faf9d1c022a1. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 760.127921] env[61987]: DEBUG oslo_concurrency.lockutils [req-b04dc650-830e-47fa-8f54-96000c4272e5 req-b30ea2ff-6a21-4df2-919b-a6a4ccb4dcf6 service nova] Acquiring lock "refresh_cache-f937a5ec-a1d7-41d8-b998-fa18f545b304" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.128043] env[61987]: DEBUG oslo_concurrency.lockutils [req-b04dc650-830e-47fa-8f54-96000c4272e5 req-b30ea2ff-6a21-4df2-919b-a6a4ccb4dcf6 service nova] Acquired lock "refresh_cache-f937a5ec-a1d7-41d8-b998-fa18f545b304" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.129023] env[61987]: DEBUG nova.network.neutron [req-b04dc650-830e-47fa-8f54-96000c4272e5 req-b30ea2ff-6a21-4df2-919b-a6a4ccb4dcf6 service nova] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Refreshing network info cache for port ac7175cc-c93a-4a76-aa42-faf9d1c022a1 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 760.208261] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquiring lock "80f13fca-2df1-4aa0-96f1-b61a60c4fdc6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.208261] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "80f13fca-2df1-4aa0-96f1-b61a60c4fdc6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.208261] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "80f13fca-2df1-4aa0-96f1-b61a60c4fdc6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.350259] env[61987]: DEBUG oslo_concurrency.lockutils [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Acquiring lock "1afcffda-0bd2-4cbe-8cfb-12a91bb50975" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.350554] env[61987]: DEBUG oslo_concurrency.lockutils [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Lock "1afcffda-0bd2-4cbe-8cfb-12a91bb50975" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.350766] env[61987]: DEBUG oslo_concurrency.lockutils [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Acquiring lock "1afcffda-0bd2-4cbe-8cfb-12a91bb50975-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.350965] env[61987]: DEBUG oslo_concurrency.lockutils [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Lock "1afcffda-0bd2-4cbe-8cfb-12a91bb50975-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.351159] env[61987]: DEBUG oslo_concurrency.lockutils [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Lock "1afcffda-0bd2-4cbe-8cfb-12a91bb50975-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.353242] env[61987]: INFO nova.compute.manager [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Terminating instance [ 760.395818] env[61987]: DEBUG oslo_vmware.api [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': task-1061543, 'name': PowerOnVM_Task, 'duration_secs': 0.86937} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.396174] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 760.396312] env[61987]: INFO nova.compute.manager [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Took 10.42 seconds to spawn the instance on the hypervisor. [ 760.396534] env[61987]: DEBUG nova.compute.manager [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 760.397603] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f07097e4-f52c-4b1f-861e-f03578fbf929 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.443892] env[61987]: DEBUG oslo_concurrency.lockutils [None req-052d393b-7e65-4114-a6f5-6f95c5796d6b tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "02fe490b-75c8-406b-98e0-aa200f249673" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.595s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.534559] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061546, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.727342] env[61987]: DEBUG nova.network.neutron [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Successfully updated port: 8a3b1e33-f1b4-48d4-b5f3-bad9dedaeb70 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 760.811296] env[61987]: DEBUG nova.compute.manager [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Stashing vm_state: active {{(pid=61987) _prep_resize /opt/stack/nova/nova/compute/manager.py:6090}} [ 760.859042] env[61987]: DEBUG oslo_concurrency.lockutils [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Acquiring lock "refresh_cache-1afcffda-0bd2-4cbe-8cfb-12a91bb50975" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.859042] env[61987]: DEBUG oslo_concurrency.lockutils [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Acquired lock "refresh_cache-1afcffda-0bd2-4cbe-8cfb-12a91bb50975" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.859465] env[61987]: DEBUG nova.network.neutron [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 760.901482] env[61987]: DEBUG nova.network.neutron [req-b04dc650-830e-47fa-8f54-96000c4272e5 req-b30ea2ff-6a21-4df2-919b-a6a4ccb4dcf6 service nova] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Updated VIF entry in instance network info cache for port ac7175cc-c93a-4a76-aa42-faf9d1c022a1. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 760.902296] env[61987]: DEBUG nova.network.neutron [req-b04dc650-830e-47fa-8f54-96000c4272e5 req-b30ea2ff-6a21-4df2-919b-a6a4ccb4dcf6 service nova] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Updating instance_info_cache with network_info: [{"id": "ac7175cc-c93a-4a76-aa42-faf9d1c022a1", "address": "fa:16:3e:4e:f6:f5", "network": {"id": "26d453d1-8edf-45e1-9e49-2a3225de6633", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1400564876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b77fbbfc7f6437083f9f30e8fdeeb35", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac7175cc-c9", "ovs_interfaceid": "ac7175cc-c93a-4a76-aa42-faf9d1c022a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.916042] env[61987]: INFO nova.compute.manager [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Took 36.50 seconds to build instance. [ 761.038859] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061546, 'name': CreateVM_Task, 'duration_secs': 0.658902} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.042123] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 761.042970] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.043172] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.043508] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 761.043818] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85466572-66ff-4fdd-bfa7-9ed71023332d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.050163] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 761.050163] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]526994c1-6581-5cb3-5c6b-5818bbd7d98c" [ 761.050163] env[61987]: _type = "Task" [ 761.050163] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.060029] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]526994c1-6581-5cb3-5c6b-5818bbd7d98c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.233644] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "refresh_cache-d566ea53-958b-4a35-989c-771180d95bb4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.233815] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquired lock "refresh_cache-d566ea53-958b-4a35-989c-771180d95bb4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.233977] env[61987]: DEBUG nova.network.neutron [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 761.288058] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquiring lock "refresh_cache-80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.288058] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquired lock "refresh_cache-80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.288257] env[61987]: DEBUG nova.network.neutron [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 761.335189] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79311b2b-943d-4438-be2d-5c0b5e0e4082 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.338884] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.344999] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684ff226-cc38-4429-b5cd-397e44c35ff2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.383331] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f70b4e8-5d91-4be2-93ef-4a7ac17c969e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.392034] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5dad271-b757-4a54-8f2c-c39d26df6e97 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.406538] env[61987]: DEBUG oslo_concurrency.lockutils [req-b04dc650-830e-47fa-8f54-96000c4272e5 req-b30ea2ff-6a21-4df2-919b-a6a4ccb4dcf6 service nova] Releasing lock "refresh_cache-f937a5ec-a1d7-41d8-b998-fa18f545b304" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.407174] env[61987]: DEBUG nova.compute.provider_tree [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.414016] env[61987]: DEBUG nova.network.neutron [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 761.418626] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0dae83b9-4969-4d49-8586-5407ffb848fc tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Lock "bb7243dc-a950-4d3e-9b17-944e57cdd56f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.639s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.540584] env[61987]: DEBUG nova.network.neutron [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.562028] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]526994c1-6581-5cb3-5c6b-5818bbd7d98c, 'name': SearchDatastore_Task, 'duration_secs': 0.038847} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.562212] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.562446] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 761.562690] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.562846] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.563088] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 761.563394] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1301ebab-75c1-410d-8442-06e98b8345ed {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.574191] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 761.574591] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 761.575434] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-358c5299-51e7-4612-8c4c-c1e46e4293dd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.582828] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 761.582828] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]526ee40a-0a26-3ed3-0332-50f5e92e22ae" [ 761.582828] env[61987]: _type = "Task" [ 761.582828] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.592202] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]526ee40a-0a26-3ed3-0332-50f5e92e22ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.792710] env[61987]: DEBUG nova.network.neutron [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 761.911838] env[61987]: DEBUG nova.scheduler.client.report [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 761.921780] env[61987]: DEBUG nova.compute.manager [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 761.974797] env[61987]: DEBUG nova.network.neutron [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Updating instance_info_cache with network_info: [{"id": "8a3b1e33-f1b4-48d4-b5f3-bad9dedaeb70", "address": "fa:16:3e:5e:2c:6b", "network": {"id": "fcbef6be-f446-46c1-ae62-cd25c2d7f5c5", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-367916424-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69749b752d1d4ad087f70f3d2ebe7b10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e547d234-640c-449b-8279-0b16f75d6627", "external-id": "nsx-vlan-transportzone-539", "segmentation_id": 539, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8a3b1e33-f1", "ovs_interfaceid": "8a3b1e33-f1b4-48d4-b5f3-bad9dedaeb70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.048315] env[61987]: DEBUG oslo_concurrency.lockutils [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Releasing lock "refresh_cache-1afcffda-0bd2-4cbe-8cfb-12a91bb50975" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.048798] env[61987]: DEBUG nova.compute.manager [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 762.049138] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 762.049901] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7998961c-a35e-4e2d-99ee-2d0f755b3c66 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.062327] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 762.062830] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7d5826b3-3c84-42cc-854f-40a583dead9b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.070647] env[61987]: DEBUG oslo_vmware.api [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for the task: (returnval){ [ 762.070647] env[61987]: value = "task-1061547" [ 762.070647] env[61987]: _type = "Task" [ 762.070647] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.084174] env[61987]: DEBUG oslo_vmware.api [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061547, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.093163] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]526ee40a-0a26-3ed3-0332-50f5e92e22ae, 'name': SearchDatastore_Task, 'duration_secs': 0.009736} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.094035] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3658d19b-56b5-43d6-a3c8-d35cb5b3d487 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.101396] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 762.101396] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]523ee842-1166-363f-537d-3be738825683" [ 762.101396] env[61987]: _type = "Task" [ 762.101396] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.109749] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523ee842-1166-363f-537d-3be738825683, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.184872] env[61987]: DEBUG nova.network.neutron [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Updating instance_info_cache with network_info: [{"id": "fc1b4b12-b396-40ac-9eab-a1309f2259e4", "address": "fa:16:3e:ec:35:ab", "network": {"id": "8a9770a4-fed3-4232-ac78-8a1fc2223229", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.118", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "705320a8b6a549d0b197ee1a35404d52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bafe8721-91d4-4127-b215-d9e8e27947dc", "external-id": "nsx-vlan-transportzone-680", "segmentation_id": 680, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc1b4b12-b3", "ovs_interfaceid": "fc1b4b12-b396-40ac-9eab-a1309f2259e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.208801] env[61987]: DEBUG nova.compute.manager [req-2d16b808-1d40-417b-97ad-a2307b549368 req-02a634f6-7798-4856-b53e-55030736d2a8 service nova] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Received event network-vif-plugged-8a3b1e33-f1b4-48d4-b5f3-bad9dedaeb70 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 762.208801] env[61987]: DEBUG oslo_concurrency.lockutils [req-2d16b808-1d40-417b-97ad-a2307b549368 req-02a634f6-7798-4856-b53e-55030736d2a8 service nova] Acquiring lock "d566ea53-958b-4a35-989c-771180d95bb4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.208801] env[61987]: DEBUG oslo_concurrency.lockutils [req-2d16b808-1d40-417b-97ad-a2307b549368 req-02a634f6-7798-4856-b53e-55030736d2a8 service nova] Lock "d566ea53-958b-4a35-989c-771180d95bb4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.208801] env[61987]: DEBUG oslo_concurrency.lockutils [req-2d16b808-1d40-417b-97ad-a2307b549368 req-02a634f6-7798-4856-b53e-55030736d2a8 service nova] Lock "d566ea53-958b-4a35-989c-771180d95bb4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.208801] env[61987]: DEBUG nova.compute.manager [req-2d16b808-1d40-417b-97ad-a2307b549368 req-02a634f6-7798-4856-b53e-55030736d2a8 service nova] [instance: d566ea53-958b-4a35-989c-771180d95bb4] No waiting events found dispatching network-vif-plugged-8a3b1e33-f1b4-48d4-b5f3-bad9dedaeb70 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 762.208801] env[61987]: WARNING nova.compute.manager [req-2d16b808-1d40-417b-97ad-a2307b549368 req-02a634f6-7798-4856-b53e-55030736d2a8 service nova] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Received unexpected event network-vif-plugged-8a3b1e33-f1b4-48d4-b5f3-bad9dedaeb70 for instance with vm_state building and task_state spawning. [ 762.209314] env[61987]: DEBUG nova.compute.manager [req-2d16b808-1d40-417b-97ad-a2307b549368 req-02a634f6-7798-4856-b53e-55030736d2a8 service nova] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Received event network-changed-8a3b1e33-f1b4-48d4-b5f3-bad9dedaeb70 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 762.209630] env[61987]: DEBUG nova.compute.manager [req-2d16b808-1d40-417b-97ad-a2307b549368 req-02a634f6-7798-4856-b53e-55030736d2a8 service nova] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Refreshing instance network info cache due to event network-changed-8a3b1e33-f1b4-48d4-b5f3-bad9dedaeb70. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 762.209989] env[61987]: DEBUG oslo_concurrency.lockutils [req-2d16b808-1d40-417b-97ad-a2307b549368 req-02a634f6-7798-4856-b53e-55030736d2a8 service nova] Acquiring lock "refresh_cache-d566ea53-958b-4a35-989c-771180d95bb4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.419941] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.511s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.420504] env[61987]: DEBUG nova.compute.manager [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 762.427019] env[61987]: DEBUG oslo_concurrency.lockutils [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.934s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.427019] env[61987]: DEBUG nova.objects.instance [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Lazy-loading 'resources' on Instance uuid 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 762.449975] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.477560] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Releasing lock "refresh_cache-d566ea53-958b-4a35-989c-771180d95bb4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.477952] env[61987]: DEBUG nova.compute.manager [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Instance network_info: |[{"id": "8a3b1e33-f1b4-48d4-b5f3-bad9dedaeb70", "address": "fa:16:3e:5e:2c:6b", "network": {"id": "fcbef6be-f446-46c1-ae62-cd25c2d7f5c5", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-367916424-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69749b752d1d4ad087f70f3d2ebe7b10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e547d234-640c-449b-8279-0b16f75d6627", "external-id": "nsx-vlan-transportzone-539", "segmentation_id": 539, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8a3b1e33-f1", "ovs_interfaceid": "8a3b1e33-f1b4-48d4-b5f3-bad9dedaeb70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 762.478364] env[61987]: DEBUG oslo_concurrency.lockutils [req-2d16b808-1d40-417b-97ad-a2307b549368 req-02a634f6-7798-4856-b53e-55030736d2a8 service nova] Acquired lock "refresh_cache-d566ea53-958b-4a35-989c-771180d95bb4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.478445] env[61987]: DEBUG nova.network.neutron [req-2d16b808-1d40-417b-97ad-a2307b549368 req-02a634f6-7798-4856-b53e-55030736d2a8 service nova] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Refreshing network info cache for port 8a3b1e33-f1b4-48d4-b5f3-bad9dedaeb70 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 762.480196] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:2c:6b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e547d234-640c-449b-8279-0b16f75d6627', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8a3b1e33-f1b4-48d4-b5f3-bad9dedaeb70', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 762.488048] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Creating folder: Project (69749b752d1d4ad087f70f3d2ebe7b10). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 762.489148] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4c2945ff-cf5a-45f5-a063-25bd8ca85d26 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.502872] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Created folder: Project (69749b752d1d4ad087f70f3d2ebe7b10) in parent group-v234219. [ 762.503191] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Creating folder: Instances. Parent ref: group-v234313. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 762.503818] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-16a5a0b6-3afb-4fce-b28f-0526b87804d7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.514136] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Created folder: Instances in parent group-v234313. [ 762.514419] env[61987]: DEBUG oslo.service.loopingcall [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 762.514635] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 762.514859] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eaf1cd19-a7a1-4253-8b09-ceaa94603420 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.535695] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 762.535695] env[61987]: value = "task-1061550" [ 762.535695] env[61987]: _type = "Task" [ 762.535695] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.544314] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061550, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.582614] env[61987]: DEBUG oslo_vmware.api [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061547, 'name': PowerOffVM_Task, 'duration_secs': 0.267026} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.582614] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 762.582614] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 762.582837] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-96df3f48-2579-4f28-8f73-06456b996f8a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.613701] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523ee842-1166-363f-537d-3be738825683, 'name': SearchDatastore_Task, 'duration_secs': 0.014014} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.615363] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.615566] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] f937a5ec-a1d7-41d8-b998-fa18f545b304/f937a5ec-a1d7-41d8-b998-fa18f545b304.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 762.615977] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 762.616209] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 762.616423] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Deleting the datastore file [datastore1] 1afcffda-0bd2-4cbe-8cfb-12a91bb50975 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 762.616829] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88d72330-8148-4aee-83d5-d1076223faa3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.619164] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0ffb473a-95f1-481e-95b3-64440e7ebe65 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.627854] env[61987]: DEBUG oslo_vmware.api [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for the task: (returnval){ [ 762.627854] env[61987]: value = "task-1061552" [ 762.627854] env[61987]: _type = "Task" [ 762.627854] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.629470] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 762.629470] env[61987]: value = "task-1061553" [ 762.629470] env[61987]: _type = "Task" [ 762.629470] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.641587] env[61987]: DEBUG oslo_vmware.api [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061552, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.644773] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061553, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.687735] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Releasing lock "refresh_cache-80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.930942] env[61987]: DEBUG nova.compute.utils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 762.932544] env[61987]: DEBUG nova.compute.manager [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 762.932973] env[61987]: DEBUG nova.network.neutron [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 762.979082] env[61987]: DEBUG nova.policy [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f9b90037d9384f549640bfcf344f0669', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '69749b752d1d4ad087f70f3d2ebe7b10', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 763.049762] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061550, 'name': CreateVM_Task, 'duration_secs': 0.391682} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.053844] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 763.054859] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.054859] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.054859] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 763.056163] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d384370-20e6-46c9-9602-13d992c6f173 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.067233] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for the task: (returnval){ [ 763.067233] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]529df0a6-69f1-09f1-df9b-ec6da80301e3" [ 763.067233] env[61987]: _type = "Task" [ 763.067233] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.091015] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]529df0a6-69f1-09f1-df9b-ec6da80301e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.145463] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061553, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.149456] env[61987]: DEBUG oslo_vmware.api [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Task: {'id': task-1061552, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12427} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.149741] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 763.149931] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 763.150805] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 763.150805] env[61987]: INFO nova.compute.manager [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Took 1.10 seconds to destroy the instance on the hypervisor. [ 763.150805] env[61987]: DEBUG oslo.service.loopingcall [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 763.151014] env[61987]: DEBUG nova.compute.manager [-] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 763.151072] env[61987]: DEBUG nova.network.neutron [-] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 763.182372] env[61987]: DEBUG nova.network.neutron [-] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 763.241531] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da4d42f6-9bc1-4d92-a320-d2791fe55a56 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.277238] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b5981d-1ed6-400e-8262-30c07f915cea {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.287220] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Updating instance '80f13fca-2df1-4aa0-96f1-b61a60c4fdc6' progress to 83 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 763.399479] env[61987]: DEBUG nova.network.neutron [req-2d16b808-1d40-417b-97ad-a2307b549368 req-02a634f6-7798-4856-b53e-55030736d2a8 service nova] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Updated VIF entry in instance network info cache for port 8a3b1e33-f1b4-48d4-b5f3-bad9dedaeb70. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 763.399479] env[61987]: DEBUG nova.network.neutron [req-2d16b808-1d40-417b-97ad-a2307b549368 req-02a634f6-7798-4856-b53e-55030736d2a8 service nova] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Updating instance_info_cache with network_info: [{"id": "8a3b1e33-f1b4-48d4-b5f3-bad9dedaeb70", "address": "fa:16:3e:5e:2c:6b", "network": {"id": "fcbef6be-f446-46c1-ae62-cd25c2d7f5c5", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-367916424-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69749b752d1d4ad087f70f3d2ebe7b10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e547d234-640c-449b-8279-0b16f75d6627", "external-id": "nsx-vlan-transportzone-539", "segmentation_id": 539, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8a3b1e33-f1", "ovs_interfaceid": "8a3b1e33-f1b4-48d4-b5f3-bad9dedaeb70", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.435841] env[61987]: DEBUG nova.compute.manager [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 763.454791] env[61987]: DEBUG nova.network.neutron [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Successfully created port: 5c0219be-ffba-47c9-a23f-e7651c879839 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 763.509497] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4afdf44f-049d-42e2-aeec-901442d2af1f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.520424] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c0d46a1-61bb-4924-8158-723190106c53 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.562510] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32635d34-a6c5-4fcb-9794-b627eb36a0e8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.572516] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52cb87f7-9171-4976-9d11-bccead254e09 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.580230] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]529df0a6-69f1-09f1-df9b-ec6da80301e3, 'name': SearchDatastore_Task, 'duration_secs': 0.054408} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.580962] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.581245] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 763.581479] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.581625] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.581806] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 763.582076] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3d0adaa2-6b27-4e86-b50e-1cbb1155387e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.591794] env[61987]: DEBUG nova.compute.provider_tree [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.607436] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 763.607638] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 763.608410] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04a5deac-0e03-47f4-aa04-d83e508edb42 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.615562] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for the task: (returnval){ [ 763.615562] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52481df7-ceba-10c7-b405-a92bc400e7d4" [ 763.615562] env[61987]: _type = "Task" [ 763.615562] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.624168] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52481df7-ceba-10c7-b405-a92bc400e7d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.645154] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061553, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.574502} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.645411] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] f937a5ec-a1d7-41d8-b998-fa18f545b304/f937a5ec-a1d7-41d8-b998-fa18f545b304.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 763.645632] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 763.645884] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e3f0d84a-f071-42ea-997e-91442ebb2f31 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.654593] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 763.654593] env[61987]: value = "task-1061554" [ 763.654593] env[61987]: _type = "Task" [ 763.654593] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.663600] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061554, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.685342] env[61987]: DEBUG nova.network.neutron [-] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.795373] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 763.795776] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66a18b0b-345f-446c-aa8e-8bff429eaf29 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.809023] env[61987]: DEBUG oslo_vmware.api [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for the task: (returnval){ [ 763.809023] env[61987]: value = "task-1061555" [ 763.809023] env[61987]: _type = "Task" [ 763.809023] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.820992] env[61987]: DEBUG oslo_vmware.api [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061555, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.902204] env[61987]: DEBUG oslo_concurrency.lockutils [req-2d16b808-1d40-417b-97ad-a2307b549368 req-02a634f6-7798-4856-b53e-55030736d2a8 service nova] Releasing lock "refresh_cache-d566ea53-958b-4a35-989c-771180d95bb4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.095580] env[61987]: DEBUG nova.scheduler.client.report [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 764.133527] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52481df7-ceba-10c7-b405-a92bc400e7d4, 'name': SearchDatastore_Task, 'duration_secs': 0.052464} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.134636] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc55918b-e24e-4bb3-a826-baab889888fa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.142811] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for the task: (returnval){ [ 764.142811] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52f780d3-0c04-cf5d-0a8e-04a433d3dcf4" [ 764.142811] env[61987]: _type = "Task" [ 764.142811] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.154743] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52f780d3-0c04-cf5d-0a8e-04a433d3dcf4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.165969] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061554, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07765} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.166355] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 764.167499] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52748231-913a-4710-9e3b-3bfff179f958 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.202509] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] f937a5ec-a1d7-41d8-b998-fa18f545b304/f937a5ec-a1d7-41d8-b998-fa18f545b304.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 764.203123] env[61987]: INFO nova.compute.manager [-] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Took 1.05 seconds to deallocate network for instance. [ 764.203841] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce8641ec-6698-4e14-be10-02b000eebca2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.230868] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 764.230868] env[61987]: value = "task-1061556" [ 764.230868] env[61987]: _type = "Task" [ 764.230868] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.241158] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061556, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.318954] env[61987]: DEBUG oslo_vmware.api [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061555, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.448919] env[61987]: DEBUG nova.compute.manager [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 764.480706] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 764.481221] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 764.481589] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 764.482113] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 764.482383] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 764.482800] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 764.483157] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 764.483372] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 764.483591] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 764.483967] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 764.484087] env[61987]: DEBUG nova.virt.hardware [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 764.485140] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f13326-ebd2-4c6a-ba87-91a9463d0cdd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.496803] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe745d8b-7835-4f5c-abe8-425489282d76 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.602135] env[61987]: DEBUG oslo_concurrency.lockutils [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.178s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.606421] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.286s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.606758] env[61987]: INFO nova.compute.claims [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 764.627734] env[61987]: INFO nova.scheduler.client.report [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Deleted allocations for instance 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b [ 764.655356] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52f780d3-0c04-cf5d-0a8e-04a433d3dcf4, 'name': SearchDatastore_Task, 'duration_secs': 0.046753} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.656461] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.656745] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] d566ea53-958b-4a35-989c-771180d95bb4/d566ea53-958b-4a35-989c-771180d95bb4.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 764.657145] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f01c0fbc-8010-49d4-a6ad-ba5021aa7553 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.665755] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for the task: (returnval){ [ 764.665755] env[61987]: value = "task-1061557" [ 764.665755] env[61987]: _type = "Task" [ 764.665755] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.680035] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061557, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.729258] env[61987]: DEBUG oslo_concurrency.lockutils [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.741358] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061556, 'name': ReconfigVM_Task, 'duration_secs': 0.481242} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.741666] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Reconfigured VM instance instance-00000036 to attach disk [datastore1] f937a5ec-a1d7-41d8-b998-fa18f545b304/f937a5ec-a1d7-41d8-b998-fa18f545b304.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 764.742340] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-245732be-3627-4b1c-966f-4e5725c1ba54 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.750047] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 764.750047] env[61987]: value = "task-1061558" [ 764.750047] env[61987]: _type = "Task" [ 764.750047] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.758850] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061558, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.818956] env[61987]: DEBUG oslo_vmware.api [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061555, 'name': PowerOnVM_Task, 'duration_secs': 0.835876} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.819331] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 764.819563] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1c56f694-c841-4d4b-8ea5-a2ff4989f50c tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Updating instance '80f13fca-2df1-4aa0-96f1-b61a60c4fdc6' progress to 100 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 765.093724] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Acquiring lock "bb7243dc-a950-4d3e-9b17-944e57cdd56f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.094160] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Lock "bb7243dc-a950-4d3e-9b17-944e57cdd56f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.094390] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Acquiring lock "bb7243dc-a950-4d3e-9b17-944e57cdd56f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.094589] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Lock "bb7243dc-a950-4d3e-9b17-944e57cdd56f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.094771] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Lock "bb7243dc-a950-4d3e-9b17-944e57cdd56f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.096971] env[61987]: INFO nova.compute.manager [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Terminating instance [ 765.140222] env[61987]: DEBUG oslo_concurrency.lockutils [None req-502727c0-3b14-4418-b53d-6bd932d31121 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680 tempest-FloatingIPsAssociationNegativeTestJSON-1426241680-project-member] Lock "8ff4ebb9-57ea-49d2-8d3c-5876da61e77b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.127s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.161367] env[61987]: DEBUG nova.compute.manager [req-b732bd76-e541-495a-b1a6-9d53c7bb104d req-24ebe4f5-8375-4695-adeb-0c6683376f41 service nova] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Received event network-vif-plugged-5c0219be-ffba-47c9-a23f-e7651c879839 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 765.161596] env[61987]: DEBUG oslo_concurrency.lockutils [req-b732bd76-e541-495a-b1a6-9d53c7bb104d req-24ebe4f5-8375-4695-adeb-0c6683376f41 service nova] Acquiring lock "8e68c5a2-6587-45b6-9a76-96b129a0a665-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.163339] env[61987]: DEBUG oslo_concurrency.lockutils [req-b732bd76-e541-495a-b1a6-9d53c7bb104d req-24ebe4f5-8375-4695-adeb-0c6683376f41 service nova] Lock "8e68c5a2-6587-45b6-9a76-96b129a0a665-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.163339] env[61987]: DEBUG oslo_concurrency.lockutils [req-b732bd76-e541-495a-b1a6-9d53c7bb104d req-24ebe4f5-8375-4695-adeb-0c6683376f41 service nova] Lock "8e68c5a2-6587-45b6-9a76-96b129a0a665-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.163339] env[61987]: DEBUG nova.compute.manager [req-b732bd76-e541-495a-b1a6-9d53c7bb104d req-24ebe4f5-8375-4695-adeb-0c6683376f41 service nova] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] No waiting events found dispatching network-vif-plugged-5c0219be-ffba-47c9-a23f-e7651c879839 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 765.163518] env[61987]: WARNING nova.compute.manager [req-b732bd76-e541-495a-b1a6-9d53c7bb104d req-24ebe4f5-8375-4695-adeb-0c6683376f41 service nova] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Received unexpected event network-vif-plugged-5c0219be-ffba-47c9-a23f-e7651c879839 for instance with vm_state building and task_state spawning. [ 765.177960] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061557, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.259705] env[61987]: DEBUG nova.network.neutron [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Successfully updated port: 5c0219be-ffba-47c9-a23f-e7651c879839 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 765.264325] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061558, 'name': Rename_Task, 'duration_secs': 0.151494} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.265095] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 765.265277] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2f9f1ae5-8886-405a-a6f8-52b217e66dc3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.275305] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 765.275305] env[61987]: value = "task-1061559" [ 765.275305] env[61987]: _type = "Task" [ 765.275305] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.286526] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061559, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.601170] env[61987]: DEBUG nova.compute.manager [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 765.601462] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 765.602842] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60df1b82-d672-4c74-97fe-2c80586a2525 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.612104] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 765.612273] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8c60c50f-3c76-4064-8b25-c1fec30054cb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.621423] env[61987]: DEBUG oslo_vmware.api [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Waiting for the task: (returnval){ [ 765.621423] env[61987]: value = "task-1061560" [ 765.621423] env[61987]: _type = "Task" [ 765.621423] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.630596] env[61987]: DEBUG oslo_vmware.api [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': task-1061560, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.682759] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061557, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.659861} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.683246] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] d566ea53-958b-4a35-989c-771180d95bb4/d566ea53-958b-4a35-989c-771180d95bb4.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 765.683403] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 765.683680] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-791fd652-39fb-4c0e-9a2d-1b6e9581a4a8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.691634] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for the task: (returnval){ [ 765.691634] env[61987]: value = "task-1061561" [ 765.691634] env[61987]: _type = "Task" [ 765.691634] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.707133] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061561, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.765508] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "refresh_cache-8e68c5a2-6587-45b6-9a76-96b129a0a665" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.765712] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquired lock "refresh_cache-8e68c5a2-6587-45b6-9a76-96b129a0a665" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.765859] env[61987]: DEBUG nova.network.neutron [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 765.797448] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061559, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.063020] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197f30f1-344c-4b5c-b1a1-33ed13e82074 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.070064] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ffe15c3-2b8f-44c4-9df0-87094a5105a0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.103336] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec68efd-9131-4ff5-8ec1-616ce2c475dc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.113082] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61a84a13-d3c7-4b98-9fe2-bd0ea813ce24 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.130407] env[61987]: DEBUG nova.compute.provider_tree [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.146115] env[61987]: DEBUG oslo_vmware.api [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': task-1061560, 'name': PowerOffVM_Task, 'duration_secs': 0.225659} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.147154] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 766.147154] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 766.147154] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1aa9ae78-a49d-425d-9bc3-d3cf13c258d7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.204518] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061561, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072264} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.204871] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 766.205700] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccaf19a5-aee7-4b7c-be56-493b7551a3fc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.229385] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] d566ea53-958b-4a35-989c-771180d95bb4/d566ea53-958b-4a35-989c-771180d95bb4.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 766.230685] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8407993d-5cff-47fa-92b4-58b05194fb69 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.247341] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 766.247341] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 766.247879] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Deleting the datastore file [datastore2] bb7243dc-a950-4d3e-9b17-944e57cdd56f {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 766.248319] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-14a1d57d-17e2-4281-bc8b-a18aa44d8835 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.258069] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for the task: (returnval){ [ 766.258069] env[61987]: value = "task-1061564" [ 766.258069] env[61987]: _type = "Task" [ 766.258069] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.260631] env[61987]: DEBUG oslo_vmware.api [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Waiting for the task: (returnval){ [ 766.260631] env[61987]: value = "task-1061563" [ 766.260631] env[61987]: _type = "Task" [ 766.260631] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.276099] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061564, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.279725] env[61987]: DEBUG oslo_vmware.api [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': task-1061563, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.289672] env[61987]: DEBUG oslo_vmware.api [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061559, 'name': PowerOnVM_Task, 'duration_secs': 0.783198} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.289985] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 766.290408] env[61987]: INFO nova.compute.manager [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Took 9.18 seconds to spawn the instance on the hypervisor. [ 766.290821] env[61987]: DEBUG nova.compute.manager [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 766.294619] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa650c2-b483-4245-aa42-8134f87451ce {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.316594] env[61987]: DEBUG nova.network.neutron [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 766.607645] env[61987]: DEBUG nova.network.neutron [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Updating instance_info_cache with network_info: [{"id": "5c0219be-ffba-47c9-a23f-e7651c879839", "address": "fa:16:3e:9e:1d:80", "network": {"id": "fcbef6be-f446-46c1-ae62-cd25c2d7f5c5", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-367916424-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69749b752d1d4ad087f70f3d2ebe7b10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e547d234-640c-449b-8279-0b16f75d6627", "external-id": "nsx-vlan-transportzone-539", "segmentation_id": 539, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c0219be-ff", "ovs_interfaceid": "5c0219be-ffba-47c9-a23f-e7651c879839", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.641177] env[61987]: DEBUG nova.scheduler.client.report [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 766.775193] env[61987]: DEBUG oslo_vmware.api [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Task: {'id': task-1061563, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147327} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.778297] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 766.778650] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 766.778905] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 766.779188] env[61987]: INFO nova.compute.manager [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Took 1.18 seconds to destroy the instance on the hypervisor. [ 766.779543] env[61987]: DEBUG oslo.service.loopingcall [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 766.779862] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061564, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.780685] env[61987]: DEBUG nova.compute.manager [-] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 766.780685] env[61987]: DEBUG nova.network.neutron [-] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 766.811166] env[61987]: INFO nova.compute.manager [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Took 34.44 seconds to build instance. [ 767.113425] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Releasing lock "refresh_cache-8e68c5a2-6587-45b6-9a76-96b129a0a665" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.113904] env[61987]: DEBUG nova.compute.manager [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Instance network_info: |[{"id": "5c0219be-ffba-47c9-a23f-e7651c879839", "address": "fa:16:3e:9e:1d:80", "network": {"id": "fcbef6be-f446-46c1-ae62-cd25c2d7f5c5", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-367916424-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69749b752d1d4ad087f70f3d2ebe7b10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e547d234-640c-449b-8279-0b16f75d6627", "external-id": "nsx-vlan-transportzone-539", "segmentation_id": 539, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c0219be-ff", "ovs_interfaceid": "5c0219be-ffba-47c9-a23f-e7651c879839", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 767.114461] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9e:1d:80', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e547d234-640c-449b-8279-0b16f75d6627', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5c0219be-ffba-47c9-a23f-e7651c879839', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 767.128304] env[61987]: DEBUG oslo.service.loopingcall [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 767.129188] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 767.129188] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a66ec462-78a1-412c-9773-e70fafbfcc43 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.148443] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.543s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.149354] env[61987]: DEBUG nova.compute.manager [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 767.154818] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.832s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.155120] env[61987]: DEBUG nova.objects.instance [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Lazy-loading 'resources' on Instance uuid 93174a10-f8b1-4789-ab3c-dda07bdddaa8 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 767.156562] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 767.156562] env[61987]: value = "task-1061565" [ 767.156562] env[61987]: _type = "Task" [ 767.156562] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.171305] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061565, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.239642] env[61987]: DEBUG nova.network.neutron [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Port fc1b4b12-b396-40ac-9eab-a1309f2259e4 binding to destination host cpu-1 is already ACTIVE {{(pid=61987) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 767.239857] env[61987]: DEBUG oslo_concurrency.lockutils [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquiring lock "refresh_cache-80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.240250] env[61987]: DEBUG oslo_concurrency.lockutils [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquired lock "refresh_cache-80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.240250] env[61987]: DEBUG nova.network.neutron [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 767.274373] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061564, 'name': ReconfigVM_Task, 'duration_secs': 0.553961} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.274939] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Reconfigured VM instance instance-00000037 to attach disk [datastore1] d566ea53-958b-4a35-989c-771180d95bb4/d566ea53-958b-4a35-989c-771180d95bb4.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 767.275729] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-454bf3b0-86da-4521-9801-9a86832c404c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.286280] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for the task: (returnval){ [ 767.286280] env[61987]: value = "task-1061566" [ 767.286280] env[61987]: _type = "Task" [ 767.286280] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.300737] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061566, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.317528] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d2a4d3c3-c40f-48e5-8c58-a70a231cfb96 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "f937a5ec-a1d7-41d8-b998-fa18f545b304" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.842s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.339142] env[61987]: DEBUG nova.compute.manager [req-a5b10149-f29d-44c7-808d-4e3911e63e42 req-f05e2829-4ab5-4e1f-9929-28f15068a93a service nova] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Received event network-changed-5c0219be-ffba-47c9-a23f-e7651c879839 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 767.339142] env[61987]: DEBUG nova.compute.manager [req-a5b10149-f29d-44c7-808d-4e3911e63e42 req-f05e2829-4ab5-4e1f-9929-28f15068a93a service nova] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Refreshing instance network info cache due to event network-changed-5c0219be-ffba-47c9-a23f-e7651c879839. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 767.339142] env[61987]: DEBUG oslo_concurrency.lockutils [req-a5b10149-f29d-44c7-808d-4e3911e63e42 req-f05e2829-4ab5-4e1f-9929-28f15068a93a service nova] Acquiring lock "refresh_cache-8e68c5a2-6587-45b6-9a76-96b129a0a665" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.339142] env[61987]: DEBUG oslo_concurrency.lockutils [req-a5b10149-f29d-44c7-808d-4e3911e63e42 req-f05e2829-4ab5-4e1f-9929-28f15068a93a service nova] Acquired lock "refresh_cache-8e68c5a2-6587-45b6-9a76-96b129a0a665" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.339142] env[61987]: DEBUG nova.network.neutron [req-a5b10149-f29d-44c7-808d-4e3911e63e42 req-f05e2829-4ab5-4e1f-9929-28f15068a93a service nova] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Refreshing network info cache for port 5c0219be-ffba-47c9-a23f-e7651c879839 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 767.667268] env[61987]: DEBUG nova.compute.utils [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 767.672709] env[61987]: DEBUG nova.compute.manager [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 767.672709] env[61987]: DEBUG nova.network.neutron [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 767.681812] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061565, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.685638] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Acquiring lock "466a5f5a-f48a-4921-8bcb-44a1840c8141" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.686317] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Lock "466a5f5a-f48a-4921-8bcb-44a1840c8141" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.686469] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Acquiring lock "466a5f5a-f48a-4921-8bcb-44a1840c8141-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.686722] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Lock "466a5f5a-f48a-4921-8bcb-44a1840c8141-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.687052] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Lock "466a5f5a-f48a-4921-8bcb-44a1840c8141-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.689095] env[61987]: INFO nova.compute.manager [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Terminating instance [ 767.748421] env[61987]: DEBUG nova.policy [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6f0a5c2d383c48989a329715c33ac26c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0f1e2e50b4144fa4a11b97b6c11c787f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 767.796339] env[61987]: DEBUG nova.network.neutron [-] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.803219] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061566, 'name': Rename_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.087589] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c84004-da5e-4941-9118-14e703906922 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.097823] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d79838c-fb7a-4bba-8b1b-bdfa0e427340 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.137984] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a6e6e26-7408-467c-87e6-afaad4a43c54 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.147142] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4809c019-033d-48dd-bfc1-4646b2eb71c9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.162112] env[61987]: DEBUG nova.compute.provider_tree [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.166068] env[61987]: DEBUG nova.network.neutron [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Successfully created port: 92111125-1a9f-426b-85f2-0450aa02f691 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 768.172483] env[61987]: DEBUG nova.compute.manager [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 768.179261] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061565, 'name': CreateVM_Task, 'duration_secs': 0.538515} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.179613] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 768.180273] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.180444] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.180767] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 768.181034] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-463cb8cc-757c-43b6-b9de-05e2265e1bcc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.188947] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for the task: (returnval){ [ 768.188947] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52ab58d9-13bc-07e3-2636-6aedcf0e6ba0" [ 768.188947] env[61987]: _type = "Task" [ 768.188947] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.193405] env[61987]: DEBUG nova.compute.manager [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 768.193645] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 768.194760] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f91b1d-47ea-4098-a8c2-b73a30aafc6c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.203529] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52ab58d9-13bc-07e3-2636-6aedcf0e6ba0, 'name': SearchDatastore_Task, 'duration_secs': 0.009911} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.205697] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.205949] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 768.206588] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.206588] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.206588] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 768.206935] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 768.207073] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8f5e3fea-0f2c-42c0-be16-bedb17640417 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.208797] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cd7ded00-07dd-466b-942f-71e8f56b141b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.217560] env[61987]: DEBUG oslo_vmware.api [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Waiting for the task: (returnval){ [ 768.217560] env[61987]: value = "task-1061567" [ 768.217560] env[61987]: _type = "Task" [ 768.217560] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.221421] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 768.221610] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 768.222620] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c60648bc-6f22-405d-a0d3-577e165db741 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.232331] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for the task: (returnval){ [ 768.232331] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]528e5b4f-7399-9183-4782-9aaa71caa36a" [ 768.232331] env[61987]: _type = "Task" [ 768.232331] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.232964] env[61987]: DEBUG oslo_vmware.api [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': task-1061567, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.242781] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]528e5b4f-7399-9183-4782-9aaa71caa36a, 'name': SearchDatastore_Task, 'duration_secs': 0.010089} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.243601] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2304e1fb-7e97-4811-b6eb-8a2456d89c16 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.250161] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for the task: (returnval){ [ 768.250161] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5280a1f5-3621-4d9a-39d7-899492b30138" [ 768.250161] env[61987]: _type = "Task" [ 768.250161] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.260200] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5280a1f5-3621-4d9a-39d7-899492b30138, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.298600] env[61987]: INFO nova.compute.manager [-] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Took 1.52 seconds to deallocate network for instance. [ 768.298935] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061566, 'name': Rename_Task, 'duration_secs': 0.871889} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.300697] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 768.303439] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9933ab9b-7055-4878-94f7-7f505452393d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.313956] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for the task: (returnval){ [ 768.313956] env[61987]: value = "task-1061568" [ 768.313956] env[61987]: _type = "Task" [ 768.313956] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.324077] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061568, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.669607] env[61987]: DEBUG nova.network.neutron [req-a5b10149-f29d-44c7-808d-4e3911e63e42 req-f05e2829-4ab5-4e1f-9929-28f15068a93a service nova] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Updated VIF entry in instance network info cache for port 5c0219be-ffba-47c9-a23f-e7651c879839. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 768.670096] env[61987]: DEBUG nova.network.neutron [req-a5b10149-f29d-44c7-808d-4e3911e63e42 req-f05e2829-4ab5-4e1f-9929-28f15068a93a service nova] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Updating instance_info_cache with network_info: [{"id": "5c0219be-ffba-47c9-a23f-e7651c879839", "address": "fa:16:3e:9e:1d:80", "network": {"id": "fcbef6be-f446-46c1-ae62-cd25c2d7f5c5", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-367916424-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69749b752d1d4ad087f70f3d2ebe7b10", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e547d234-640c-449b-8279-0b16f75d6627", "external-id": "nsx-vlan-transportzone-539", "segmentation_id": 539, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c0219be-ff", "ovs_interfaceid": "5c0219be-ffba-47c9-a23f-e7651c879839", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.673812] env[61987]: DEBUG nova.scheduler.client.report [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 768.690176] env[61987]: DEBUG nova.network.neutron [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Updating instance_info_cache with network_info: [{"id": "fc1b4b12-b396-40ac-9eab-a1309f2259e4", "address": "fa:16:3e:ec:35:ab", "network": {"id": "8a9770a4-fed3-4232-ac78-8a1fc2223229", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.118", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "705320a8b6a549d0b197ee1a35404d52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bafe8721-91d4-4127-b215-d9e8e27947dc", "external-id": "nsx-vlan-transportzone-680", "segmentation_id": 680, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc1b4b12-b3", "ovs_interfaceid": "fc1b4b12-b396-40ac-9eab-a1309f2259e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.729746] env[61987]: DEBUG oslo_vmware.api [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': task-1061567, 'name': PowerOffVM_Task, 'duration_secs': 0.298495} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.730301] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 768.730640] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 768.731560] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0a79c366-bd8c-4f96-9377-f88c991f97a9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.761633] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5280a1f5-3621-4d9a-39d7-899492b30138, 'name': SearchDatastore_Task, 'duration_secs': 0.009952} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.762045] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.762297] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 8e68c5a2-6587-45b6-9a76-96b129a0a665/8e68c5a2-6587-45b6-9a76-96b129a0a665.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 768.762578] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d52cd23f-7a64-43e7-a437-d8b0fc0275d8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.771559] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for the task: (returnval){ [ 768.771559] env[61987]: value = "task-1061570" [ 768.771559] env[61987]: _type = "Task" [ 768.771559] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.780667] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061570, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.806863] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.819209] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 768.819209] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 768.819209] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Deleting the datastore file [datastore2] 466a5f5a-f48a-4921-8bcb-44a1840c8141 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 768.823400] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b3ec50f5-895b-4c3f-863d-192c5022d8d2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.832921] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061568, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.838451] env[61987]: DEBUG nova.compute.manager [req-8394b4c6-ce5d-4975-9f6b-d241d6b21107 req-ddd674df-27fe-46ac-b3ba-83e24c4553b6 service nova] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Received event network-changed-ac7175cc-c93a-4a76-aa42-faf9d1c022a1 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 768.838451] env[61987]: DEBUG nova.compute.manager [req-8394b4c6-ce5d-4975-9f6b-d241d6b21107 req-ddd674df-27fe-46ac-b3ba-83e24c4553b6 service nova] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Refreshing instance network info cache due to event network-changed-ac7175cc-c93a-4a76-aa42-faf9d1c022a1. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 768.838451] env[61987]: DEBUG oslo_concurrency.lockutils [req-8394b4c6-ce5d-4975-9f6b-d241d6b21107 req-ddd674df-27fe-46ac-b3ba-83e24c4553b6 service nova] Acquiring lock "refresh_cache-f937a5ec-a1d7-41d8-b998-fa18f545b304" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.838451] env[61987]: DEBUG oslo_concurrency.lockutils [req-8394b4c6-ce5d-4975-9f6b-d241d6b21107 req-ddd674df-27fe-46ac-b3ba-83e24c4553b6 service nova] Acquired lock "refresh_cache-f937a5ec-a1d7-41d8-b998-fa18f545b304" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.838451] env[61987]: DEBUG nova.network.neutron [req-8394b4c6-ce5d-4975-9f6b-d241d6b21107 req-ddd674df-27fe-46ac-b3ba-83e24c4553b6 service nova] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Refreshing network info cache for port ac7175cc-c93a-4a76-aa42-faf9d1c022a1 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 768.840543] env[61987]: DEBUG oslo_vmware.api [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Waiting for the task: (returnval){ [ 768.840543] env[61987]: value = "task-1061571" [ 768.840543] env[61987]: _type = "Task" [ 768.840543] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.856162] env[61987]: DEBUG oslo_vmware.api [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': task-1061571, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.182601] env[61987]: DEBUG oslo_concurrency.lockutils [req-a5b10149-f29d-44c7-808d-4e3911e63e42 req-f05e2829-4ab5-4e1f-9929-28f15068a93a service nova] Releasing lock "refresh_cache-8e68c5a2-6587-45b6-9a76-96b129a0a665" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.194709] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.040s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.198682] env[61987]: DEBUG nova.compute.manager [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 769.201810] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.008s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.202145] env[61987]: DEBUG nova.objects.instance [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Lazy-loading 'resources' on Instance uuid d55395cc-0b92-4e99-9b80-913c19f6c7f4 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 769.204426] env[61987]: DEBUG oslo_concurrency.lockutils [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Releasing lock "refresh_cache-80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.235696] env[61987]: DEBUG nova.virt.hardware [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 769.236768] env[61987]: DEBUG nova.virt.hardware [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 769.236931] env[61987]: DEBUG nova.virt.hardware [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 769.237437] env[61987]: DEBUG nova.virt.hardware [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 769.237663] env[61987]: DEBUG nova.virt.hardware [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 769.237880] env[61987]: DEBUG nova.virt.hardware [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 769.238185] env[61987]: DEBUG nova.virt.hardware [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 769.238324] env[61987]: DEBUG nova.virt.hardware [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 769.238514] env[61987]: DEBUG nova.virt.hardware [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 769.238713] env[61987]: DEBUG nova.virt.hardware [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 769.239028] env[61987]: DEBUG nova.virt.hardware [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 769.240523] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14ebb6a3-6dae-436b-8387-104b4414cd15 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.245326] env[61987]: INFO nova.scheduler.client.report [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Deleted allocations for instance 93174a10-f8b1-4789-ab3c-dda07bdddaa8 [ 769.256157] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac800064-22df-47ef-add6-ddc493394f18 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.288013] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061570, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.505134} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.288767] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 8e68c5a2-6587-45b6-9a76-96b129a0a665/8e68c5a2-6587-45b6-9a76-96b129a0a665.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 769.288767] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 769.288943] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0b7cb9a6-93c2-4017-8c4e-fe3b2bcf2817 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.298306] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for the task: (returnval){ [ 769.298306] env[61987]: value = "task-1061572" [ 769.298306] env[61987]: _type = "Task" [ 769.298306] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.307810] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061572, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.329120] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061568, 'name': PowerOnVM_Task, 'duration_secs': 0.621357} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.329410] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 769.329620] env[61987]: INFO nova.compute.manager [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Took 9.61 seconds to spawn the instance on the hypervisor. [ 769.330216] env[61987]: DEBUG nova.compute.manager [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 769.330706] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6020eb00-c3a8-4c8d-af68-dde1b1bd84cf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.355440] env[61987]: DEBUG oslo_vmware.api [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Task: {'id': task-1061571, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.245666} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.355777] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 769.355976] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 769.356238] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 769.356544] env[61987]: INFO nova.compute.manager [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Took 1.16 seconds to destroy the instance on the hypervisor. [ 769.356855] env[61987]: DEBUG oslo.service.loopingcall [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 769.357094] env[61987]: DEBUG nova.compute.manager [-] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 769.357186] env[61987]: DEBUG nova.network.neutron [-] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 769.413432] env[61987]: DEBUG nova.compute.manager [req-dbafee70-eac2-41bc-9977-5b40c89ba6c4 req-efbb141b-840e-4def-8763-3ff1ccb3e24b service nova] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Received event network-vif-deleted-04dbba0f-63b4-4d42-a3ec-7fe8dbd70488 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 769.712073] env[61987]: DEBUG nova.compute.manager [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61987) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:900}} [ 769.712325] env[61987]: DEBUG oslo_concurrency.lockutils [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.759941] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4af570ef-a453-4db2-8477-7d8c958ef540 tempest-ServersTestMultiNic-1197793481 tempest-ServersTestMultiNic-1197793481-project-member] Lock "93174a10-f8b1-4789-ab3c-dda07bdddaa8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.037s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.813452] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061572, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069091} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.815366] env[61987]: DEBUG nova.network.neutron [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Successfully updated port: 92111125-1a9f-426b-85f2-0450aa02f691 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 769.816353] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 769.817424] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c3d2b47-d255-4661-be2e-153377a81368 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.847745] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] 8e68c5a2-6587-45b6-9a76-96b129a0a665/8e68c5a2-6587-45b6-9a76-96b129a0a665.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 769.854714] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6166591a-61a5-4006-adc3-083b6f2cdbb8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.872561] env[61987]: DEBUG nova.network.neutron [req-8394b4c6-ce5d-4975-9f6b-d241d6b21107 req-ddd674df-27fe-46ac-b3ba-83e24c4553b6 service nova] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Updated VIF entry in instance network info cache for port ac7175cc-c93a-4a76-aa42-faf9d1c022a1. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 769.872916] env[61987]: DEBUG nova.network.neutron [req-8394b4c6-ce5d-4975-9f6b-d241d6b21107 req-ddd674df-27fe-46ac-b3ba-83e24c4553b6 service nova] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Updating instance_info_cache with network_info: [{"id": "ac7175cc-c93a-4a76-aa42-faf9d1c022a1", "address": "fa:16:3e:4e:f6:f5", "network": {"id": "26d453d1-8edf-45e1-9e49-2a3225de6633", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1400564876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.167", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b77fbbfc7f6437083f9f30e8fdeeb35", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac7175cc-c9", "ovs_interfaceid": "ac7175cc-c93a-4a76-aa42-faf9d1c022a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.874733] env[61987]: INFO nova.compute.manager [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Took 36.55 seconds to build instance. [ 769.884029] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for the task: (returnval){ [ 769.884029] env[61987]: value = "task-1061573" [ 769.884029] env[61987]: _type = "Task" [ 769.884029] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.901021] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061573, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.178289] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac674b6-4f26-4ef9-b086-9cc1fec71798 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.188299] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5022238e-6ebe-4239-9b0b-4e875ea00049 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.221731] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6da8c10-83ad-4a88-b478-7bbb6b9cec39 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.230225] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09400468-6ad9-4028-9431-8fa3ff49c402 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.247531] env[61987]: DEBUG nova.compute.provider_tree [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.318304] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Acquiring lock "refresh_cache-1c63bf62-3f70-42b9-b6d7-41336d20f6c8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.318455] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Acquired lock "refresh_cache-1c63bf62-3f70-42b9-b6d7-41336d20f6c8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.319045] env[61987]: DEBUG nova.network.neutron [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 770.376222] env[61987]: DEBUG oslo_concurrency.lockutils [req-8394b4c6-ce5d-4975-9f6b-d241d6b21107 req-ddd674df-27fe-46ac-b3ba-83e24c4553b6 service nova] Releasing lock "refresh_cache-f937a5ec-a1d7-41d8-b998-fa18f545b304" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.376724] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "d566ea53-958b-4a35-989c-771180d95bb4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.208s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.395413] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061573, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.609456] env[61987]: DEBUG nova.network.neutron [-] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.752395] env[61987]: DEBUG nova.scheduler.client.report [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 770.862668] env[61987]: DEBUG nova.compute.manager [req-8295358a-6150-4c54-b580-4f48918db3d1 req-0af21670-9e1a-439b-8bf6-a831c6c45daa service nova] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Received event network-vif-deleted-39a5fbe3-c262-40f4-acdc-48209a7f617a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 770.898206] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061573, 'name': ReconfigVM_Task, 'duration_secs': 0.716167} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.898516] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Reconfigured VM instance instance-00000038 to attach disk [datastore1] 8e68c5a2-6587-45b6-9a76-96b129a0a665/8e68c5a2-6587-45b6-9a76-96b129a0a665.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 770.899138] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2c1277d9-8f3d-4b0c-a6c2-071d7ee8844c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.904469] env[61987]: DEBUG nova.network.neutron [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 770.908031] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for the task: (returnval){ [ 770.908031] env[61987]: value = "task-1061574" [ 770.908031] env[61987]: _type = "Task" [ 770.908031] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.916962] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061574, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.115695] env[61987]: INFO nova.compute.manager [-] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Took 1.76 seconds to deallocate network for instance. [ 771.147954] env[61987]: DEBUG nova.network.neutron [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Updating instance_info_cache with network_info: [{"id": "92111125-1a9f-426b-85f2-0450aa02f691", "address": "fa:16:3e:c3:4c:f9", "network": {"id": "b9a9fb0c-f607-4035-8099-ce8318968e2d", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-2001138182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f1e2e50b4144fa4a11b97b6c11c787f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92111125-1a", "ovs_interfaceid": "92111125-1a9f-426b-85f2-0450aa02f691", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.258369] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.056s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.262350] env[61987]: DEBUG oslo_concurrency.lockutils [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.612s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.263135] env[61987]: INFO nova.compute.claims [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 771.297088] env[61987]: INFO nova.scheduler.client.report [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Deleted allocations for instance d55395cc-0b92-4e99-9b80-913c19f6c7f4 [ 771.424819] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061574, 'name': Rename_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.466022] env[61987]: DEBUG nova.compute.manager [req-7b3ab572-5bc3-46d6-a86d-57bee71e0bef req-9016e7c4-e002-482d-8f44-0276b620b085 service nova] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Received event network-vif-plugged-92111125-1a9f-426b-85f2-0450aa02f691 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 771.466022] env[61987]: DEBUG oslo_concurrency.lockutils [req-7b3ab572-5bc3-46d6-a86d-57bee71e0bef req-9016e7c4-e002-482d-8f44-0276b620b085 service nova] Acquiring lock "1c63bf62-3f70-42b9-b6d7-41336d20f6c8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.466022] env[61987]: DEBUG oslo_concurrency.lockutils [req-7b3ab572-5bc3-46d6-a86d-57bee71e0bef req-9016e7c4-e002-482d-8f44-0276b620b085 service nova] Lock "1c63bf62-3f70-42b9-b6d7-41336d20f6c8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.466022] env[61987]: DEBUG oslo_concurrency.lockutils [req-7b3ab572-5bc3-46d6-a86d-57bee71e0bef req-9016e7c4-e002-482d-8f44-0276b620b085 service nova] Lock "1c63bf62-3f70-42b9-b6d7-41336d20f6c8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.466022] env[61987]: DEBUG nova.compute.manager [req-7b3ab572-5bc3-46d6-a86d-57bee71e0bef req-9016e7c4-e002-482d-8f44-0276b620b085 service nova] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] No waiting events found dispatching network-vif-plugged-92111125-1a9f-426b-85f2-0450aa02f691 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 771.466022] env[61987]: WARNING nova.compute.manager [req-7b3ab572-5bc3-46d6-a86d-57bee71e0bef req-9016e7c4-e002-482d-8f44-0276b620b085 service nova] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Received unexpected event network-vif-plugged-92111125-1a9f-426b-85f2-0450aa02f691 for instance with vm_state building and task_state spawning. [ 771.466022] env[61987]: DEBUG nova.compute.manager [req-7b3ab572-5bc3-46d6-a86d-57bee71e0bef req-9016e7c4-e002-482d-8f44-0276b620b085 service nova] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Received event network-changed-92111125-1a9f-426b-85f2-0450aa02f691 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 771.466022] env[61987]: DEBUG nova.compute.manager [req-7b3ab572-5bc3-46d6-a86d-57bee71e0bef req-9016e7c4-e002-482d-8f44-0276b620b085 service nova] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Refreshing instance network info cache due to event network-changed-92111125-1a9f-426b-85f2-0450aa02f691. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 771.466783] env[61987]: DEBUG oslo_concurrency.lockutils [req-7b3ab572-5bc3-46d6-a86d-57bee71e0bef req-9016e7c4-e002-482d-8f44-0276b620b085 service nova] Acquiring lock "refresh_cache-1c63bf62-3f70-42b9-b6d7-41336d20f6c8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.632777] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.653429] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Releasing lock "refresh_cache-1c63bf62-3f70-42b9-b6d7-41336d20f6c8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.653820] env[61987]: DEBUG nova.compute.manager [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Instance network_info: |[{"id": "92111125-1a9f-426b-85f2-0450aa02f691", "address": "fa:16:3e:c3:4c:f9", "network": {"id": "b9a9fb0c-f607-4035-8099-ce8318968e2d", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-2001138182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f1e2e50b4144fa4a11b97b6c11c787f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92111125-1a", "ovs_interfaceid": "92111125-1a9f-426b-85f2-0450aa02f691", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 771.654174] env[61987]: DEBUG oslo_concurrency.lockutils [req-7b3ab572-5bc3-46d6-a86d-57bee71e0bef req-9016e7c4-e002-482d-8f44-0276b620b085 service nova] Acquired lock "refresh_cache-1c63bf62-3f70-42b9-b6d7-41336d20f6c8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.654548] env[61987]: DEBUG nova.network.neutron [req-7b3ab572-5bc3-46d6-a86d-57bee71e0bef req-9016e7c4-e002-482d-8f44-0276b620b085 service nova] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Refreshing network info cache for port 92111125-1a9f-426b-85f2-0450aa02f691 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 771.655772] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:4c:f9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2be3fdb5-359e-43bd-8c20-2ff00e81db55', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '92111125-1a9f-426b-85f2-0450aa02f691', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 771.668062] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Creating folder: Project (0f1e2e50b4144fa4a11b97b6c11c787f). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 771.673079] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5136fb19-e8a0-4fdf-ad7d-4e2c8f867cca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.686691] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Created folder: Project (0f1e2e50b4144fa4a11b97b6c11c787f) in parent group-v234219. [ 771.686912] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Creating folder: Instances. Parent ref: group-v234317. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 771.687185] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ba96dcc8-cac6-41af-b0c2-e0a32b8be9b3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.695825] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Created folder: Instances in parent group-v234317. [ 771.696315] env[61987]: DEBUG oslo.service.loopingcall [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.696315] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 771.696770] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-082f0b57-0544-42ac-b981-e650b4355b29 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.719610] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 771.719610] env[61987]: value = "task-1061577" [ 771.719610] env[61987]: _type = "Task" [ 771.719610] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.731662] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061577, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.808520] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f0170c58-8354-432c-adcb-76671eb245bf tempest-ServerTagsTestJSON-437829592 tempest-ServerTagsTestJSON-437829592-project-member] Lock "d55395cc-0b92-4e99-9b80-913c19f6c7f4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.334s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.921476] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061574, 'name': Rename_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.115344] env[61987]: DEBUG nova.network.neutron [req-7b3ab572-5bc3-46d6-a86d-57bee71e0bef req-9016e7c4-e002-482d-8f44-0276b620b085 service nova] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Updated VIF entry in instance network info cache for port 92111125-1a9f-426b-85f2-0450aa02f691. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 772.115703] env[61987]: DEBUG nova.network.neutron [req-7b3ab572-5bc3-46d6-a86d-57bee71e0bef req-9016e7c4-e002-482d-8f44-0276b620b085 service nova] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Updating instance_info_cache with network_info: [{"id": "92111125-1a9f-426b-85f2-0450aa02f691", "address": "fa:16:3e:c3:4c:f9", "network": {"id": "b9a9fb0c-f607-4035-8099-ce8318968e2d", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-2001138182-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0f1e2e50b4144fa4a11b97b6c11c787f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2be3fdb5-359e-43bd-8c20-2ff00e81db55", "external-id": "nsx-vlan-transportzone-986", "segmentation_id": 986, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92111125-1a", "ovs_interfaceid": "92111125-1a9f-426b-85f2-0450aa02f691", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.230959] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061577, 'name': CreateVM_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.424040] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061574, 'name': Rename_Task, 'duration_secs': 1.20241} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.424377] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 772.424641] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-69402aa6-da00-40b9-bcb9-69f7582ec1f0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.430793] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for the task: (returnval){ [ 772.430793] env[61987]: value = "task-1061578" [ 772.430793] env[61987]: _type = "Task" [ 772.430793] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.441707] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061578, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.620052] env[61987]: DEBUG oslo_concurrency.lockutils [req-7b3ab572-5bc3-46d6-a86d-57bee71e0bef req-9016e7c4-e002-482d-8f44-0276b620b085 service nova] Releasing lock "refresh_cache-1c63bf62-3f70-42b9-b6d7-41336d20f6c8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.645105] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67fecc60-b541-49f5-8795-fff80f23c44c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.653180] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b865638d-e571-42ec-a8d2-a7a615bb0bcf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.686511] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ac2ddf-62e6-4a26-82a1-5380881d7deb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.694892] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7128e360-57b5-4d5a-b082-eeb6c0cdebd0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.709911] env[61987]: DEBUG nova.compute.provider_tree [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.731328] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061577, 'name': CreateVM_Task, 'duration_secs': 0.793382} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.731626] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 772.732432] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.732772] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.733208] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 772.733557] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59ae62ef-6a4b-4f6d-ac81-3c2ca20f509c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.738092] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Waiting for the task: (returnval){ [ 772.738092] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]526212f0-0414-26b7-7717-5876447b3ca8" [ 772.738092] env[61987]: _type = "Task" [ 772.738092] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.746117] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]526212f0-0414-26b7-7717-5876447b3ca8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.942225] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061578, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.216744] env[61987]: DEBUG nova.scheduler.client.report [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 773.257296] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]526212f0-0414-26b7-7717-5876447b3ca8, 'name': SearchDatastore_Task, 'duration_secs': 0.014615} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.257962] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.258656] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 773.259416] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.259810] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.260412] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 773.263257] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1d0bd09a-99ef-43e2-ad29-7cb84951bf17 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.270240] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 773.270505] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 773.271538] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0a6e1dc-759c-46ab-bd14-c63c342d9449 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.281938] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Waiting for the task: (returnval){ [ 773.281938] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52ece099-4719-6c62-1cb0-1274914e145b" [ 773.281938] env[61987]: _type = "Task" [ 773.281938] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.293092] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52ece099-4719-6c62-1cb0-1274914e145b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.444212] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061578, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.722861] env[61987]: DEBUG oslo_concurrency.lockutils [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.461s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.723444] env[61987]: DEBUG nova.compute.manager [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 773.726513] env[61987]: DEBUG oslo_concurrency.lockutils [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.542s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.728900] env[61987]: INFO nova.compute.claims [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 773.794418] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52ece099-4719-6c62-1cb0-1274914e145b, 'name': SearchDatastore_Task, 'duration_secs': 0.009342} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.794827] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b0758d8-9f76-4809-a5fc-a5c235819a3a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.801378] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Waiting for the task: (returnval){ [ 773.801378] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]526be667-b6c6-9701-6b7d-99443c054cba" [ 773.801378] env[61987]: _type = "Task" [ 773.801378] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.809244] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]526be667-b6c6-9701-6b7d-99443c054cba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.945109] env[61987]: DEBUG oslo_vmware.api [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061578, 'name': PowerOnVM_Task, 'duration_secs': 1.159294} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.945427] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 773.945766] env[61987]: INFO nova.compute.manager [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Took 9.50 seconds to spawn the instance on the hypervisor. [ 773.946072] env[61987]: DEBUG nova.compute.manager [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 773.946954] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32608706-f267-4dd7-9ab1-98e0c34c7168 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.233119] env[61987]: DEBUG nova.compute.utils [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 774.234300] env[61987]: DEBUG nova.compute.manager [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 774.234433] env[61987]: DEBUG nova.network.neutron [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 774.312969] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]526be667-b6c6-9701-6b7d-99443c054cba, 'name': SearchDatastore_Task, 'duration_secs': 0.010537} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.313288] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.313562] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 1c63bf62-3f70-42b9-b6d7-41336d20f6c8/1c63bf62-3f70-42b9-b6d7-41336d20f6c8.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 774.313863] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-09f008e3-57db-4fa2-8dd2-43d07f989d58 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.317576] env[61987]: DEBUG nova.policy [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9073b17b601445fa9e4eb979debc2f4d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3fd98395e57149c987f82585296751b2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 774.324240] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Waiting for the task: (returnval){ [ 774.324240] env[61987]: value = "task-1061579" [ 774.324240] env[61987]: _type = "Task" [ 774.324240] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.333101] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Task: {'id': task-1061579, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.471911] env[61987]: INFO nova.compute.manager [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Took 40.46 seconds to build instance. [ 774.738269] env[61987]: DEBUG nova.compute.manager [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 774.840351] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Task: {'id': task-1061579, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509174} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.840351] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 1c63bf62-3f70-42b9-b6d7-41336d20f6c8/1c63bf62-3f70-42b9-b6d7-41336d20f6c8.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 774.840351] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 774.840351] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-49d69ea6-57ac-4069-a1ac-5db90dfe951b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.849197] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Waiting for the task: (returnval){ [ 774.849197] env[61987]: value = "task-1061580" [ 774.849197] env[61987]: _type = "Task" [ 774.849197] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.861024] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Task: {'id': task-1061580, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.948835] env[61987]: DEBUG nova.network.neutron [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Successfully created port: 7eb7297c-3eaf-4825-b148-c0e6a54ec595 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 774.974860] env[61987]: DEBUG oslo_concurrency.lockutils [None req-91875f91-807b-4c5d-9afd-386172b3025b tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "8e68c5a2-6587-45b6-9a76-96b129a0a665" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 104.768s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.249254] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f4c41b2-8a92-4f69-ab46-e518071eff27 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.258462] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1505183e-4c11-4406-81db-5653ce77afb8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.292119] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b5507a1-e212-4cc9-ad5d-1218d9fe5d7d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.300700] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c9eebd3-2fd2-463c-b250-3457d6ecc649 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.321684] env[61987]: DEBUG nova.compute.provider_tree [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 775.357691] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Task: {'id': task-1061580, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070182} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.357992] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 775.358788] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-621e0074-18e3-42f0-8f2f-0dc2cd8c3e1f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.388597] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 1c63bf62-3f70-42b9-b6d7-41336d20f6c8/1c63bf62-3f70-42b9-b6d7-41336d20f6c8.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 775.388911] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-366fc6bb-b9b1-4854-8ecc-d7a7f883909d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.417109] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Waiting for the task: (returnval){ [ 775.417109] env[61987]: value = "task-1061581" [ 775.417109] env[61987]: _type = "Task" [ 775.417109] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.425062] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Task: {'id': task-1061581, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.754404] env[61987]: DEBUG nova.compute.manager [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 775.786537] env[61987]: DEBUG nova.virt.hardware [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 775.786819] env[61987]: DEBUG nova.virt.hardware [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 775.786985] env[61987]: DEBUG nova.virt.hardware [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 775.787352] env[61987]: DEBUG nova.virt.hardware [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 775.787584] env[61987]: DEBUG nova.virt.hardware [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 775.787790] env[61987]: DEBUG nova.virt.hardware [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 775.788299] env[61987]: DEBUG nova.virt.hardware [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 775.788500] env[61987]: DEBUG nova.virt.hardware [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 775.788972] env[61987]: DEBUG nova.virt.hardware [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 775.788972] env[61987]: DEBUG nova.virt.hardware [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 775.789099] env[61987]: DEBUG nova.virt.hardware [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 775.789935] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b92dae9b-1e00-4c6d-87ff-d06b4562c696 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.799388] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cfa2b5a-81ec-4162-96fc-400fd8d5e67d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.824623] env[61987]: DEBUG nova.scheduler.client.report [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 775.931157] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Task: {'id': task-1061581, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.934786] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "d566ea53-958b-4a35-989c-771180d95bb4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.934786] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "d566ea53-958b-4a35-989c-771180d95bb4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.934786] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "d566ea53-958b-4a35-989c-771180d95bb4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.934786] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "d566ea53-958b-4a35-989c-771180d95bb4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.934786] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "d566ea53-958b-4a35-989c-771180d95bb4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.936060] env[61987]: INFO nova.compute.manager [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Terminating instance [ 776.004465] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Acquiring lock "59d34772-26a5-4af2-82dc-da932e996745" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.004710] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Lock "59d34772-26a5-4af2-82dc-da932e996745" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.121016] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "8e68c5a2-6587-45b6-9a76-96b129a0a665" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.121340] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "8e68c5a2-6587-45b6-9a76-96b129a0a665" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.121559] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "8e68c5a2-6587-45b6-9a76-96b129a0a665-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.121753] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "8e68c5a2-6587-45b6-9a76-96b129a0a665-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.121987] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "8e68c5a2-6587-45b6-9a76-96b129a0a665-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.124252] env[61987]: INFO nova.compute.manager [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Terminating instance [ 776.332292] env[61987]: DEBUG oslo_concurrency.lockutils [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.606s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.333572] env[61987]: DEBUG nova.compute.manager [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 776.337729] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 33.537s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.338193] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.338390] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61987) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 776.338672] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.582s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.340973] env[61987]: INFO nova.compute.claims [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 776.346709] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5445f36d-9c7e-473a-9bf6-9e536ac8f800 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.361190] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06eaa53d-6234-41ba-884c-132c199b791d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.377154] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aa6a4e4-766f-429c-bd90-ed2ce9ce27d9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.386338] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bcf5be1-f11f-40cc-a318-e4cee4b86202 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.423873] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181439MB free_disk=179GB free_vcpus=48 pci_devices=None {{(pid=61987) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 776.424038] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.433256] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Task: {'id': task-1061581, 'name': ReconfigVM_Task, 'duration_secs': 0.823345} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.433539] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 1c63bf62-3f70-42b9-b6d7-41336d20f6c8/1c63bf62-3f70-42b9-b6d7-41336d20f6c8.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 776.434361] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-45fc9625-2b78-41bf-bec3-88876cbbad0f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.441081] env[61987]: DEBUG nova.compute.manager [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 776.441310] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 776.441740] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Waiting for the task: (returnval){ [ 776.441740] env[61987]: value = "task-1061582" [ 776.441740] env[61987]: _type = "Task" [ 776.441740] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.442473] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6d417c0-c1a4-4dd4-96fb-018c0f5f783d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.453326] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 776.456461] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-124f1afb-6d25-4b04-8511-df6c558a5914 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.458051] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Task: {'id': task-1061582, 'name': Rename_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.462834] env[61987]: DEBUG oslo_vmware.api [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for the task: (returnval){ [ 776.462834] env[61987]: value = "task-1061583" [ 776.462834] env[61987]: _type = "Task" [ 776.462834] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.472271] env[61987]: DEBUG oslo_vmware.api [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061583, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.508322] env[61987]: DEBUG nova.compute.manager [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 776.628774] env[61987]: DEBUG nova.compute.manager [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 776.628999] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 776.630599] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e76cdce-e4bb-4b1d-b10d-024c1a3c7b7a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.639916] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 776.639916] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5cccaaa3-91ef-472e-aca4-d4bc63bfcf40 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.647046] env[61987]: DEBUG oslo_vmware.api [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for the task: (returnval){ [ 776.647046] env[61987]: value = "task-1061584" [ 776.647046] env[61987]: _type = "Task" [ 776.647046] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.658453] env[61987]: DEBUG oslo_vmware.api [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061584, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.851127] env[61987]: DEBUG nova.compute.utils [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 776.851655] env[61987]: DEBUG nova.compute.manager [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 776.852054] env[61987]: DEBUG nova.network.neutron [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 776.959657] env[61987]: DEBUG nova.policy [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3011f049be54b2cb8922d646b94a310', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd828262de7b14d2ab38193d7d34e8f7a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 776.961267] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Task: {'id': task-1061582, 'name': Rename_Task, 'duration_secs': 0.157298} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.961595] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 776.962279] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0e24d221-c794-405d-b4a4-31b67086d49d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.972134] env[61987]: DEBUG oslo_vmware.api [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061583, 'name': PowerOffVM_Task, 'duration_secs': 0.202378} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.973328] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 776.973515] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 776.973834] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Waiting for the task: (returnval){ [ 776.973834] env[61987]: value = "task-1061585" [ 776.973834] env[61987]: _type = "Task" [ 776.973834] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.974027] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-303fc064-c207-4e71-a022-e9e6cef381c5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.983545] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Task: {'id': task-1061585, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.033290] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.036655] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 777.036655] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 777.036812] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Deleting the datastore file [datastore1] d566ea53-958b-4a35-989c-771180d95bb4 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 777.036932] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ba3595f4-f7c9-4a1f-938b-e9e2483f6523 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.043692] env[61987]: DEBUG oslo_vmware.api [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for the task: (returnval){ [ 777.043692] env[61987]: value = "task-1061587" [ 777.043692] env[61987]: _type = "Task" [ 777.043692] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.053837] env[61987]: DEBUG oslo_vmware.api [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061587, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.087621] env[61987]: DEBUG nova.compute.manager [req-f275f4e0-e215-450e-84fb-fbc8a8a5a52d req-d0473183-b3ec-4188-8aa9-7c40d499736e service nova] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Received event network-vif-plugged-7eb7297c-3eaf-4825-b148-c0e6a54ec595 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 777.087863] env[61987]: DEBUG oslo_concurrency.lockutils [req-f275f4e0-e215-450e-84fb-fbc8a8a5a52d req-d0473183-b3ec-4188-8aa9-7c40d499736e service nova] Acquiring lock "93a27dea-f28a-4cbf-ac37-ff41763ae18d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.088111] env[61987]: DEBUG oslo_concurrency.lockutils [req-f275f4e0-e215-450e-84fb-fbc8a8a5a52d req-d0473183-b3ec-4188-8aa9-7c40d499736e service nova] Lock "93a27dea-f28a-4cbf-ac37-ff41763ae18d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.088298] env[61987]: DEBUG oslo_concurrency.lockutils [req-f275f4e0-e215-450e-84fb-fbc8a8a5a52d req-d0473183-b3ec-4188-8aa9-7c40d499736e service nova] Lock "93a27dea-f28a-4cbf-ac37-ff41763ae18d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.088478] env[61987]: DEBUG nova.compute.manager [req-f275f4e0-e215-450e-84fb-fbc8a8a5a52d req-d0473183-b3ec-4188-8aa9-7c40d499736e service nova] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] No waiting events found dispatching network-vif-plugged-7eb7297c-3eaf-4825-b148-c0e6a54ec595 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 777.088652] env[61987]: WARNING nova.compute.manager [req-f275f4e0-e215-450e-84fb-fbc8a8a5a52d req-d0473183-b3ec-4188-8aa9-7c40d499736e service nova] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Received unexpected event network-vif-plugged-7eb7297c-3eaf-4825-b148-c0e6a54ec595 for instance with vm_state building and task_state spawning. [ 777.158508] env[61987]: DEBUG oslo_vmware.api [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061584, 'name': PowerOffVM_Task, 'duration_secs': 0.164846} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.158792] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 777.158986] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 777.159294] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-86ad332e-fb17-4c59-a48e-cdd568bfe051 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.239853] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 777.240225] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 777.240400] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Deleting the datastore file [datastore1] 8e68c5a2-6587-45b6-9a76-96b129a0a665 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 777.240694] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7124981a-2bc3-4e4b-8562-51ad4eb25ca8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.247814] env[61987]: DEBUG oslo_vmware.api [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for the task: (returnval){ [ 777.247814] env[61987]: value = "task-1061589" [ 777.247814] env[61987]: _type = "Task" [ 777.247814] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.256246] env[61987]: DEBUG oslo_vmware.api [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061589, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.352886] env[61987]: DEBUG nova.compute.manager [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 777.413025] env[61987]: DEBUG nova.network.neutron [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Successfully updated port: 7eb7297c-3eaf-4825-b148-c0e6a54ec595 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 777.491188] env[61987]: DEBUG oslo_vmware.api [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Task: {'id': task-1061585, 'name': PowerOnVM_Task, 'duration_secs': 0.473057} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.491460] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 777.491692] env[61987]: INFO nova.compute.manager [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Took 8.29 seconds to spawn the instance on the hypervisor. [ 777.491912] env[61987]: DEBUG nova.compute.manager [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 777.492968] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd7f830-dd9b-4d79-9251-8f75d7114d8d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.565096] env[61987]: DEBUG oslo_vmware.api [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061587, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155586} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.565497] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 777.565755] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 777.566029] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 777.566280] env[61987]: INFO nova.compute.manager [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Took 1.12 seconds to destroy the instance on the hypervisor. [ 777.570017] env[61987]: DEBUG oslo.service.loopingcall [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 777.570017] env[61987]: DEBUG nova.compute.manager [-] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 777.570017] env[61987]: DEBUG nova.network.neutron [-] [instance: d566ea53-958b-4a35-989c-771180d95bb4] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 777.641168] env[61987]: DEBUG nova.network.neutron [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Successfully created port: 979ec444-341c-4296-b3b7-060800602dd1 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 777.758975] env[61987]: DEBUG oslo_vmware.api [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Task: {'id': task-1061589, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143773} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.759199] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 777.759639] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 777.759639] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 777.759757] env[61987]: INFO nova.compute.manager [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Took 1.13 seconds to destroy the instance on the hypervisor. [ 777.760013] env[61987]: DEBUG oslo.service.loopingcall [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 777.760207] env[61987]: DEBUG nova.compute.manager [-] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 777.760420] env[61987]: DEBUG nova.network.neutron [-] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 777.791715] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-069e64ca-6bdf-4eb4-9d9a-0b3055aa6a7e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.799359] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd0c39a1-b359-46ed-b3dc-0d5927597cf9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.833113] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquiring lock "d038391c-2b32-455c-85d5-68221e7bd065" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.833400] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Lock "d038391c-2b32-455c-85d5-68221e7bd065" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.833611] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquiring lock "d038391c-2b32-455c-85d5-68221e7bd065-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.833797] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Lock "d038391c-2b32-455c-85d5-68221e7bd065-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.834046] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Lock "d038391c-2b32-455c-85d5-68221e7bd065-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.837280] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a439bf5-8d19-438b-9683-13360fba2d26 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.840734] env[61987]: INFO nova.compute.manager [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Terminating instance [ 777.848469] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68eef2ee-b579-4c6c-aacf-bd024582a7eb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.866704] env[61987]: DEBUG nova.compute.provider_tree [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 777.916858] env[61987]: DEBUG oslo_concurrency.lockutils [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "refresh_cache-93a27dea-f28a-4cbf-ac37-ff41763ae18d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.917027] env[61987]: DEBUG oslo_concurrency.lockutils [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquired lock "refresh_cache-93a27dea-f28a-4cbf-ac37-ff41763ae18d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.917188] env[61987]: DEBUG nova.network.neutron [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 778.019746] env[61987]: INFO nova.compute.manager [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Took 40.73 seconds to build instance. [ 778.336928] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Acquiring lock "fe22fc1b-61fb-4088-b5fe-b84eb4ec4099" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.337143] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Lock "fe22fc1b-61fb-4088-b5fe-b84eb4ec4099" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.345870] env[61987]: DEBUG nova.compute.manager [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 778.346897] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 778.348304] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b87c97-d24b-4bf0-9bb8-da021b3a794f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.361155] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 778.361378] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3815289d-6edb-45aa-aac9-0b534b648d95 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.369568] env[61987]: DEBUG nova.compute.manager [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 778.374013] env[61987]: DEBUG oslo_vmware.api [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 778.374013] env[61987]: value = "task-1061590" [ 778.374013] env[61987]: _type = "Task" [ 778.374013] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.374013] env[61987]: DEBUG nova.scheduler.client.report [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 778.388990] env[61987]: DEBUG oslo_vmware.api [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061590, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.414919] env[61987]: DEBUG nova.virt.hardware [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 778.414919] env[61987]: DEBUG nova.virt.hardware [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 778.415070] env[61987]: DEBUG nova.virt.hardware [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 778.415333] env[61987]: DEBUG nova.virt.hardware [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 778.415333] env[61987]: DEBUG nova.virt.hardware [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 778.415456] env[61987]: DEBUG nova.virt.hardware [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 778.415668] env[61987]: DEBUG nova.virt.hardware [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 778.416645] env[61987]: DEBUG nova.virt.hardware [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 778.416645] env[61987]: DEBUG nova.virt.hardware [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 778.416645] env[61987]: DEBUG nova.virt.hardware [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 778.416645] env[61987]: DEBUG nova.virt.hardware [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 778.417929] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd72196-e28f-4b80-896b-84c45b02c9c7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.427831] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7091e3ed-e119-4b56-950e-8750bd466455 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.487206] env[61987]: DEBUG nova.network.neutron [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 778.521625] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c33de706-d488-4bd6-907e-7696caecf3c7 tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Lock "1c63bf62-3f70-42b9-b6d7-41336d20f6c8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.353s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.609048] env[61987]: DEBUG nova.network.neutron [-] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.722773] env[61987]: DEBUG nova.network.neutron [-] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.804427] env[61987]: DEBUG nova.network.neutron [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Updating instance_info_cache with network_info: [{"id": "7eb7297c-3eaf-4825-b148-c0e6a54ec595", "address": "fa:16:3e:52:1e:02", "network": {"id": "abcd2857-4c5a-4546-a003-858ddb17b0e6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-124838733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3fd98395e57149c987f82585296751b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7eb7297c-3e", "ovs_interfaceid": "7eb7297c-3eaf-4825-b148-c0e6a54ec595", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.840783] env[61987]: DEBUG nova.compute.manager [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 778.883403] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.545s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.883985] env[61987]: DEBUG nova.compute.manager [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 778.886983] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.307s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.890322] env[61987]: DEBUG nova.objects.instance [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Lazy-loading 'resources' on Instance uuid 630e0959-6d7f-401d-b84c-1c098f54bf60 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 778.897946] env[61987]: DEBUG oslo_vmware.api [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061590, 'name': PowerOffVM_Task, 'duration_secs': 0.237904} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.898214] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 778.898392] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 778.899289] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-efd851e4-2d16-47b7-a12b-309fb1a75312 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.932901] env[61987]: DEBUG nova.compute.manager [req-a06fe223-f1fd-430f-90db-12df5b0612fa req-8316d006-69c0-4b4d-85d5-af14d42be27a service nova] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Received event network-vif-deleted-5c0219be-ffba-47c9-a23f-e7651c879839 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 778.974902] env[61987]: DEBUG oslo_concurrency.lockutils [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquiring lock "93706a54-1fc4-4aed-8807-ec2c3d3694b4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.975190] env[61987]: DEBUG oslo_concurrency.lockutils [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Lock "93706a54-1fc4-4aed-8807-ec2c3d3694b4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.978292] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 778.978402] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 778.978596] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Deleting the datastore file [datastore2] d038391c-2b32-455c-85d5-68221e7bd065 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 778.979043] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-190a23fd-db33-48fb-b266-56d318041456 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.987983] env[61987]: DEBUG oslo_vmware.api [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 778.987983] env[61987]: value = "task-1061592" [ 778.987983] env[61987]: _type = "Task" [ 778.987983] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.998405] env[61987]: DEBUG oslo_vmware.api [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061592, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.111870] env[61987]: INFO nova.compute.manager [-] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Took 1.54 seconds to deallocate network for instance. [ 779.212477] env[61987]: DEBUG nova.compute.manager [req-e59bafc2-9256-4ec4-b7d4-437e7c4e0f39 req-7fa028c9-eb40-4a2f-ba6d-2b9c957de95a service nova] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Received event network-changed-7eb7297c-3eaf-4825-b148-c0e6a54ec595 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 779.212580] env[61987]: DEBUG nova.compute.manager [req-e59bafc2-9256-4ec4-b7d4-437e7c4e0f39 req-7fa028c9-eb40-4a2f-ba6d-2b9c957de95a service nova] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Refreshing instance network info cache due to event network-changed-7eb7297c-3eaf-4825-b148-c0e6a54ec595. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 779.212724] env[61987]: DEBUG oslo_concurrency.lockutils [req-e59bafc2-9256-4ec4-b7d4-437e7c4e0f39 req-7fa028c9-eb40-4a2f-ba6d-2b9c957de95a service nova] Acquiring lock "refresh_cache-93a27dea-f28a-4cbf-ac37-ff41763ae18d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.222976] env[61987]: INFO nova.compute.manager [-] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Took 1.46 seconds to deallocate network for instance. [ 779.310686] env[61987]: DEBUG oslo_concurrency.lockutils [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Releasing lock "refresh_cache-93a27dea-f28a-4cbf-ac37-ff41763ae18d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.310686] env[61987]: DEBUG nova.compute.manager [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Instance network_info: |[{"id": "7eb7297c-3eaf-4825-b148-c0e6a54ec595", "address": "fa:16:3e:52:1e:02", "network": {"id": "abcd2857-4c5a-4546-a003-858ddb17b0e6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-124838733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3fd98395e57149c987f82585296751b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7eb7297c-3e", "ovs_interfaceid": "7eb7297c-3eaf-4825-b148-c0e6a54ec595", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 779.310686] env[61987]: DEBUG oslo_concurrency.lockutils [req-e59bafc2-9256-4ec4-b7d4-437e7c4e0f39 req-7fa028c9-eb40-4a2f-ba6d-2b9c957de95a service nova] Acquired lock "refresh_cache-93a27dea-f28a-4cbf-ac37-ff41763ae18d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.310686] env[61987]: DEBUG nova.network.neutron [req-e59bafc2-9256-4ec4-b7d4-437e7c4e0f39 req-7fa028c9-eb40-4a2f-ba6d-2b9c957de95a service nova] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Refreshing network info cache for port 7eb7297c-3eaf-4825-b148-c0e6a54ec595 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 779.312891] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:1e:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '170f3b82-5915-4e36-bce9-4664ebb6be5e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7eb7297c-3eaf-4825-b148-c0e6a54ec595', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 779.322574] env[61987]: DEBUG oslo.service.loopingcall [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 779.323601] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 779.324040] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8f4b6824-c813-430e-afdb-95d3580fed08 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.346998] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 779.346998] env[61987]: value = "task-1061593" [ 779.346998] env[61987]: _type = "Task" [ 779.346998] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.359098] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061593, 'name': CreateVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.366603] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.397683] env[61987]: DEBUG nova.compute.utils [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 779.400372] env[61987]: DEBUG nova.compute.manager [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 779.400372] env[61987]: DEBUG nova.network.neutron [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 779.462048] env[61987]: DEBUG nova.policy [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7d15929a29cb498485c2b3bf77da9457', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d8a522ba9bf422db98ca69c8d92b373', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 779.482116] env[61987]: DEBUG nova.compute.manager [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 779.500739] env[61987]: DEBUG oslo_vmware.api [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061592, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138434} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.501012] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 779.502409] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 779.502409] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 779.502409] env[61987]: INFO nova.compute.manager [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Took 1.16 seconds to destroy the instance on the hypervisor. [ 779.502409] env[61987]: DEBUG oslo.service.loopingcall [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 779.502409] env[61987]: DEBUG nova.compute.manager [-] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 779.502409] env[61987]: DEBUG nova.network.neutron [-] [instance: d038391c-2b32-455c-85d5-68221e7bd065] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 779.620991] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.713326] env[61987]: DEBUG nova.network.neutron [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Successfully updated port: 979ec444-341c-4296-b3b7-060800602dd1 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 779.731339] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.861621] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061593, 'name': CreateVM_Task, 'duration_secs': 0.469919} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.861810] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 779.862732] env[61987]: DEBUG oslo_concurrency.lockutils [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.863381] env[61987]: DEBUG oslo_concurrency.lockutils [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.863488] env[61987]: DEBUG oslo_concurrency.lockutils [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 779.864097] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa05e9ef-8654-4e85-933d-ae81f6082e97 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.871388] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 779.871388] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52f979cf-f164-83ba-a337-43a4da6079ce" [ 779.871388] env[61987]: _type = "Task" [ 779.871388] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.881318] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52f979cf-f164-83ba-a337-43a4da6079ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.895979] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9d2d60-76a6-4cdf-98f2-09c18588f8ab {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.904053] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aef8d83c-9793-4b38-9f89-04366fb7b5a0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.908131] env[61987]: DEBUG nova.compute.manager [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 779.911695] env[61987]: DEBUG nova.network.neutron [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Successfully created port: b9f8f891-82e4-440d-bf4d-5b3c70c910e6 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 779.944282] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be184e4e-75be-4f80-b5cf-5679c4e87bf6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.955144] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5585083f-0c80-48bc-83ba-552354ca046a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.971225] env[61987]: DEBUG nova.compute.provider_tree [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.994195] env[61987]: DEBUG oslo_concurrency.lockutils [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Acquiring lock "1c63bf62-3f70-42b9-b6d7-41336d20f6c8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.994474] env[61987]: DEBUG oslo_concurrency.lockutils [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Lock "1c63bf62-3f70-42b9-b6d7-41336d20f6c8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.994888] env[61987]: DEBUG oslo_concurrency.lockutils [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Acquiring lock "1c63bf62-3f70-42b9-b6d7-41336d20f6c8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.994888] env[61987]: DEBUG oslo_concurrency.lockutils [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Lock "1c63bf62-3f70-42b9-b6d7-41336d20f6c8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.995143] env[61987]: DEBUG oslo_concurrency.lockutils [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Lock "1c63bf62-3f70-42b9-b6d7-41336d20f6c8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.997554] env[61987]: INFO nova.compute.manager [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Terminating instance [ 780.012279] env[61987]: DEBUG oslo_concurrency.lockutils [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.220111] env[61987]: DEBUG oslo_concurrency.lockutils [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "refresh_cache-30c4fff0-c77e-4c11-aced-a040003a6b10" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.220111] env[61987]: DEBUG oslo_concurrency.lockutils [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquired lock "refresh_cache-30c4fff0-c77e-4c11-aced-a040003a6b10" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.220111] env[61987]: DEBUG nova.network.neutron [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 780.307213] env[61987]: DEBUG nova.network.neutron [req-e59bafc2-9256-4ec4-b7d4-437e7c4e0f39 req-7fa028c9-eb40-4a2f-ba6d-2b9c957de95a service nova] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Updated VIF entry in instance network info cache for port 7eb7297c-3eaf-4825-b148-c0e6a54ec595. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 780.307213] env[61987]: DEBUG nova.network.neutron [req-e59bafc2-9256-4ec4-b7d4-437e7c4e0f39 req-7fa028c9-eb40-4a2f-ba6d-2b9c957de95a service nova] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Updating instance_info_cache with network_info: [{"id": "7eb7297c-3eaf-4825-b148-c0e6a54ec595", "address": "fa:16:3e:52:1e:02", "network": {"id": "abcd2857-4c5a-4546-a003-858ddb17b0e6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-124838733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3fd98395e57149c987f82585296751b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7eb7297c-3e", "ovs_interfaceid": "7eb7297c-3eaf-4825-b148-c0e6a54ec595", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.384748] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52f979cf-f164-83ba-a337-43a4da6079ce, 'name': SearchDatastore_Task, 'duration_secs': 0.009532} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.385064] env[61987]: DEBUG oslo_concurrency.lockutils [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.385310] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 780.385556] env[61987]: DEBUG oslo_concurrency.lockutils [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.385707] env[61987]: DEBUG oslo_concurrency.lockutils [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.385890] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 780.386201] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c1923015-d940-489b-a897-48c1f84cacdb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.397994] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 780.397994] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 780.397994] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-adf7a12e-2abb-43b4-8697-4754613a2d9e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.401508] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 780.401508] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52f2e7f7-6c6e-8d4e-d4cb-09b13b96e7fd" [ 780.401508] env[61987]: _type = "Task" [ 780.401508] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.409824] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52f2e7f7-6c6e-8d4e-d4cb-09b13b96e7fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.420731] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Acquiring lock "98d895ee-1d1f-49cd-a055-fe6c8462a692" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.420974] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Lock "98d895ee-1d1f-49cd-a055-fe6c8462a692" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.473861] env[61987]: DEBUG nova.scheduler.client.report [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 780.501120] env[61987]: DEBUG nova.compute.manager [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 780.501393] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 780.506022] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dffb01c3-ca6d-414c-8f66-27ce7dec9fbb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.512657] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 780.513082] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d6648358-db5d-4ed6-ac21-cfdcf0b5cc44 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.520666] env[61987]: DEBUG oslo_vmware.api [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Waiting for the task: (returnval){ [ 780.520666] env[61987]: value = "task-1061594" [ 780.520666] env[61987]: _type = "Task" [ 780.520666] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.532081] env[61987]: DEBUG oslo_vmware.api [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Task: {'id': task-1061594, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.755505] env[61987]: DEBUG nova.network.neutron [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 780.803226] env[61987]: DEBUG nova.network.neutron [-] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.813942] env[61987]: DEBUG oslo_concurrency.lockutils [req-e59bafc2-9256-4ec4-b7d4-437e7c4e0f39 req-7fa028c9-eb40-4a2f-ba6d-2b9c957de95a service nova] Releasing lock "refresh_cache-93a27dea-f28a-4cbf-ac37-ff41763ae18d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.813942] env[61987]: DEBUG nova.compute.manager [req-e59bafc2-9256-4ec4-b7d4-437e7c4e0f39 req-7fa028c9-eb40-4a2f-ba6d-2b9c957de95a service nova] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Received event network-vif-deleted-8a3b1e33-f1b4-48d4-b5f3-bad9dedaeb70 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 780.920789] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52f2e7f7-6c6e-8d4e-d4cb-09b13b96e7fd, 'name': SearchDatastore_Task, 'duration_secs': 0.008976} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.928022] env[61987]: DEBUG nova.compute.manager [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 780.928643] env[61987]: DEBUG nova.compute.manager [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 780.931210] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a951d5c5-febf-4fb7-a2a1-fbb11f9899ab {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.938189] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 780.938189] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]524e0655-f748-3227-1056-eec32fc078f2" [ 780.938189] env[61987]: _type = "Task" [ 780.938189] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.947625] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524e0655-f748-3227-1056-eec32fc078f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.979179] env[61987]: DEBUG nova.virt.hardware [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 780.979611] env[61987]: DEBUG nova.virt.hardware [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 780.979782] env[61987]: DEBUG nova.virt.hardware [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 780.979875] env[61987]: DEBUG nova.virt.hardware [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 780.980045] env[61987]: DEBUG nova.virt.hardware [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 780.981303] env[61987]: DEBUG nova.virt.hardware [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 780.981303] env[61987]: DEBUG nova.virt.hardware [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 780.981303] env[61987]: DEBUG nova.virt.hardware [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 780.981303] env[61987]: DEBUG nova.virt.hardware [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 780.981303] env[61987]: DEBUG nova.virt.hardware [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 780.981303] env[61987]: DEBUG nova.virt.hardware [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 780.982493] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.095s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.984598] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1780fb15-6fd6-4102-a9a8-0be9f9d0a5d2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.988932] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 32.432s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.988932] env[61987]: DEBUG nova.objects.instance [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61987) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 780.998103] env[61987]: DEBUG nova.network.neutron [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Updating instance_info_cache with network_info: [{"id": "979ec444-341c-4296-b3b7-060800602dd1", "address": "fa:16:3e:ad:c4:72", "network": {"id": "e2d0ce90-ff14-47de-bc4e-a2564d1268a7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-222691221-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d828262de7b14d2ab38193d7d34e8f7a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap979ec444-34", "ovs_interfaceid": "979ec444-341c-4296-b3b7-060800602dd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.004615] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-772e1716-338e-4791-8c39-f90223c5c3f4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.020421] env[61987]: INFO nova.scheduler.client.report [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Deleted allocations for instance 630e0959-6d7f-401d-b84c-1c098f54bf60 [ 781.034875] env[61987]: DEBUG oslo_vmware.api [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Task: {'id': task-1061594, 'name': PowerOffVM_Task, 'duration_secs': 0.258863} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.035197] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 781.035373] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 781.036201] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-40290313-0b92-4f85-b3da-c54afe479a08 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.101924] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 781.102188] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 781.102379] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Deleting the datastore file [datastore1] 1c63bf62-3f70-42b9-b6d7-41336d20f6c8 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 781.102645] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b5f2c065-95e8-4754-93aa-9266d60339f1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.112676] env[61987]: DEBUG oslo_vmware.api [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Waiting for the task: (returnval){ [ 781.112676] env[61987]: value = "task-1061596" [ 781.112676] env[61987]: _type = "Task" [ 781.112676] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.122624] env[61987]: DEBUG oslo_vmware.api [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Task: {'id': task-1061596, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.240339] env[61987]: DEBUG nova.compute.manager [req-027e1d0a-c554-477f-b014-6c5ff96f1519 req-24c6143d-12ab-4650-b76e-7ff013a3928d service nova] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Received event network-vif-plugged-979ec444-341c-4296-b3b7-060800602dd1 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 781.240565] env[61987]: DEBUG oslo_concurrency.lockutils [req-027e1d0a-c554-477f-b014-6c5ff96f1519 req-24c6143d-12ab-4650-b76e-7ff013a3928d service nova] Acquiring lock "30c4fff0-c77e-4c11-aced-a040003a6b10-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.240839] env[61987]: DEBUG oslo_concurrency.lockutils [req-027e1d0a-c554-477f-b014-6c5ff96f1519 req-24c6143d-12ab-4650-b76e-7ff013a3928d service nova] Lock "30c4fff0-c77e-4c11-aced-a040003a6b10-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.241074] env[61987]: DEBUG oslo_concurrency.lockutils [req-027e1d0a-c554-477f-b014-6c5ff96f1519 req-24c6143d-12ab-4650-b76e-7ff013a3928d service nova] Lock "30c4fff0-c77e-4c11-aced-a040003a6b10-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.241291] env[61987]: DEBUG nova.compute.manager [req-027e1d0a-c554-477f-b014-6c5ff96f1519 req-24c6143d-12ab-4650-b76e-7ff013a3928d service nova] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] No waiting events found dispatching network-vif-plugged-979ec444-341c-4296-b3b7-060800602dd1 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 781.241506] env[61987]: WARNING nova.compute.manager [req-027e1d0a-c554-477f-b014-6c5ff96f1519 req-24c6143d-12ab-4650-b76e-7ff013a3928d service nova] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Received unexpected event network-vif-plugged-979ec444-341c-4296-b3b7-060800602dd1 for instance with vm_state building and task_state spawning. [ 781.241717] env[61987]: DEBUG nova.compute.manager [req-027e1d0a-c554-477f-b014-6c5ff96f1519 req-24c6143d-12ab-4650-b76e-7ff013a3928d service nova] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Received event network-changed-979ec444-341c-4296-b3b7-060800602dd1 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 781.241912] env[61987]: DEBUG nova.compute.manager [req-027e1d0a-c554-477f-b014-6c5ff96f1519 req-24c6143d-12ab-4650-b76e-7ff013a3928d service nova] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Refreshing instance network info cache due to event network-changed-979ec444-341c-4296-b3b7-060800602dd1. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 781.242101] env[61987]: DEBUG oslo_concurrency.lockutils [req-027e1d0a-c554-477f-b014-6c5ff96f1519 req-24c6143d-12ab-4650-b76e-7ff013a3928d service nova] Acquiring lock "refresh_cache-30c4fff0-c77e-4c11-aced-a040003a6b10" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.306909] env[61987]: INFO nova.compute.manager [-] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Took 1.80 seconds to deallocate network for instance. [ 781.450071] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524e0655-f748-3227-1056-eec32fc078f2, 'name': SearchDatastore_Task, 'duration_secs': 0.009845} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.450366] env[61987]: DEBUG oslo_concurrency.lockutils [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.450630] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 93a27dea-f28a-4cbf-ac37-ff41763ae18d/93a27dea-f28a-4cbf-ac37-ff41763ae18d.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 781.450889] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-daaba077-8f43-4b7d-838e-6e523c18c799 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.453613] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.458216] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 781.458216] env[61987]: value = "task-1061597" [ 781.458216] env[61987]: _type = "Task" [ 781.458216] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.465830] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061597, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.492014] env[61987]: DEBUG nova.compute.manager [req-775ed773-d914-4f81-a6d6-a42ffb7b89f6 req-9d1bfbe3-3780-4f47-a0e3-af1ef74b0950 service nova] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Received event network-vif-plugged-b9f8f891-82e4-440d-bf4d-5b3c70c910e6 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 781.492324] env[61987]: DEBUG oslo_concurrency.lockutils [req-775ed773-d914-4f81-a6d6-a42ffb7b89f6 req-9d1bfbe3-3780-4f47-a0e3-af1ef74b0950 service nova] Acquiring lock "10354ba0-5a64-4174-9c61-152d54a5dfcc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.492519] env[61987]: DEBUG oslo_concurrency.lockutils [req-775ed773-d914-4f81-a6d6-a42ffb7b89f6 req-9d1bfbe3-3780-4f47-a0e3-af1ef74b0950 service nova] Lock "10354ba0-5a64-4174-9c61-152d54a5dfcc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.492706] env[61987]: DEBUG oslo_concurrency.lockutils [req-775ed773-d914-4f81-a6d6-a42ffb7b89f6 req-9d1bfbe3-3780-4f47-a0e3-af1ef74b0950 service nova] Lock "10354ba0-5a64-4174-9c61-152d54a5dfcc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.493535] env[61987]: DEBUG nova.compute.manager [req-775ed773-d914-4f81-a6d6-a42ffb7b89f6 req-9d1bfbe3-3780-4f47-a0e3-af1ef74b0950 service nova] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] No waiting events found dispatching network-vif-plugged-b9f8f891-82e4-440d-bf4d-5b3c70c910e6 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 781.493763] env[61987]: WARNING nova.compute.manager [req-775ed773-d914-4f81-a6d6-a42ffb7b89f6 req-9d1bfbe3-3780-4f47-a0e3-af1ef74b0950 service nova] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Received unexpected event network-vif-plugged-b9f8f891-82e4-440d-bf4d-5b3c70c910e6 for instance with vm_state building and task_state spawning. [ 781.502821] env[61987]: DEBUG oslo_concurrency.lockutils [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Releasing lock "refresh_cache-30c4fff0-c77e-4c11-aced-a040003a6b10" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.502948] env[61987]: DEBUG nova.compute.manager [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Instance network_info: |[{"id": "979ec444-341c-4296-b3b7-060800602dd1", "address": "fa:16:3e:ad:c4:72", "network": {"id": "e2d0ce90-ff14-47de-bc4e-a2564d1268a7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-222691221-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d828262de7b14d2ab38193d7d34e8f7a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap979ec444-34", "ovs_interfaceid": "979ec444-341c-4296-b3b7-060800602dd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 781.503225] env[61987]: DEBUG oslo_concurrency.lockutils [req-027e1d0a-c554-477f-b014-6c5ff96f1519 req-24c6143d-12ab-4650-b76e-7ff013a3928d service nova] Acquired lock "refresh_cache-30c4fff0-c77e-4c11-aced-a040003a6b10" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.503430] env[61987]: DEBUG nova.network.neutron [req-027e1d0a-c554-477f-b014-6c5ff96f1519 req-24c6143d-12ab-4650-b76e-7ff013a3928d service nova] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Refreshing network info cache for port 979ec444-341c-4296-b3b7-060800602dd1 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 781.504656] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ad:c4:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f52a458-d157-48a3-b4e2-b8cc0779afe2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '979ec444-341c-4296-b3b7-060800602dd1', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 781.512748] env[61987]: DEBUG oslo.service.loopingcall [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 781.516438] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 781.517324] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-43327f3a-76b7-4ac7-8c68-b81f0e5f2e40 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.536761] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a9ec2b9e-8f18-45ca-8677-385031e82c63 tempest-ImagesOneServerTestJSON-1683633474 tempest-ImagesOneServerTestJSON-1683633474-project-member] Lock "630e0959-6d7f-401d-b84c-1c098f54bf60" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.004s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.541195] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 781.541195] env[61987]: value = "task-1061598" [ 781.541195] env[61987]: _type = "Task" [ 781.541195] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.549239] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061598, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.622982] env[61987]: DEBUG oslo_vmware.api [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Task: {'id': task-1061596, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169999} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.623274] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 781.624171] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 781.624171] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 781.624171] env[61987]: INFO nova.compute.manager [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Took 1.12 seconds to destroy the instance on the hypervisor. [ 781.624171] env[61987]: DEBUG oslo.service.loopingcall [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 781.627027] env[61987]: DEBUG nova.compute.manager [-] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 781.627133] env[61987]: DEBUG nova.network.neutron [-] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 781.639432] env[61987]: DEBUG nova.network.neutron [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Successfully updated port: b9f8f891-82e4-440d-bf4d-5b3c70c910e6 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 781.817565] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.971240] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061597, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.972097] env[61987]: DEBUG nova.network.neutron [req-027e1d0a-c554-477f-b014-6c5ff96f1519 req-24c6143d-12ab-4650-b76e-7ff013a3928d service nova] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Updated VIF entry in instance network info cache for port 979ec444-341c-4296-b3b7-060800602dd1. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 781.972507] env[61987]: DEBUG nova.network.neutron [req-027e1d0a-c554-477f-b014-6c5ff96f1519 req-24c6143d-12ab-4650-b76e-7ff013a3928d service nova] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Updating instance_info_cache with network_info: [{"id": "979ec444-341c-4296-b3b7-060800602dd1", "address": "fa:16:3e:ad:c4:72", "network": {"id": "e2d0ce90-ff14-47de-bc4e-a2564d1268a7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-222691221-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d828262de7b14d2ab38193d7d34e8f7a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap979ec444-34", "ovs_interfaceid": "979ec444-341c-4296-b3b7-060800602dd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.001536] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c63eb30f-9596-483f-8cbb-c08beb95dd66 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.004502] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.158s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.006232] env[61987]: INFO nova.compute.claims [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 782.054391] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061598, 'name': CreateVM_Task, 'duration_secs': 0.456028} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.054778] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 782.055616] env[61987]: DEBUG oslo_concurrency.lockutils [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.056048] env[61987]: DEBUG oslo_concurrency.lockutils [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.056543] env[61987]: DEBUG oslo_concurrency.lockutils [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 782.056933] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4583432-831b-4885-88fb-ab92d34e68ad {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.065126] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 782.065126] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5255080f-902b-6a21-e552-d31c5186cd23" [ 782.065126] env[61987]: _type = "Task" [ 782.065126] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.075856] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5255080f-902b-6a21-e552-d31c5186cd23, 'name': SearchDatastore_Task, 'duration_secs': 0.010274} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.076719] env[61987]: DEBUG oslo_concurrency.lockutils [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.076719] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 782.076964] env[61987]: DEBUG oslo_concurrency.lockutils [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.077119] env[61987]: DEBUG oslo_concurrency.lockutils [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.077350] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 782.078133] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-275ca548-0387-4c8d-b821-42094dc5b2c1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.086126] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 782.086358] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 782.087232] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0be64f0-0776-4481-a452-98e67f7e58db {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.092481] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 782.092481] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52744252-4f5b-6b1f-a376-dc6a4ec70b78" [ 782.092481] env[61987]: _type = "Task" [ 782.092481] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.101774] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52744252-4f5b-6b1f-a376-dc6a4ec70b78, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.123870] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "ad40937f-d5a9-4708-8b81-06087d38d765" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.124142] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ad40937f-d5a9-4708-8b81-06087d38d765" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.141556] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquiring lock "refresh_cache-10354ba0-5a64-4174-9c61-152d54a5dfcc" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.141705] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquired lock "refresh_cache-10354ba0-5a64-4174-9c61-152d54a5dfcc" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.141874] env[61987]: DEBUG nova.network.neutron [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 782.471814] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061597, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.514434} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.472102] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 93a27dea-f28a-4cbf-ac37-ff41763ae18d/93a27dea-f28a-4cbf-ac37-ff41763ae18d.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 782.472356] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 782.472595] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-db2ee96a-bf5b-43f0-97d6-65f9f0c09587 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.475391] env[61987]: DEBUG oslo_concurrency.lockutils [req-027e1d0a-c554-477f-b014-6c5ff96f1519 req-24c6143d-12ab-4650-b76e-7ff013a3928d service nova] Releasing lock "refresh_cache-30c4fff0-c77e-4c11-aced-a040003a6b10" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.475643] env[61987]: DEBUG nova.compute.manager [req-027e1d0a-c554-477f-b014-6c5ff96f1519 req-24c6143d-12ab-4650-b76e-7ff013a3928d service nova] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Received event network-vif-deleted-203d9e60-9c1f-4c51-9ee8-9a34be19d289 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 782.479992] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 782.479992] env[61987]: value = "task-1061599" [ 782.479992] env[61987]: _type = "Task" [ 782.479992] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.488543] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061599, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.610125] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52744252-4f5b-6b1f-a376-dc6a4ec70b78, 'name': SearchDatastore_Task, 'duration_secs': 0.008181} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.610977] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ecb5e66-ed29-4e80-a36b-c4a7e03cdda2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.617562] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 782.617562] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52c7ad67-ac9f-5ff8-0cea-1bdfdb74bc9f" [ 782.617562] env[61987]: _type = "Task" [ 782.617562] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.630459] env[61987]: DEBUG nova.compute.utils [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 782.630459] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c7ad67-ac9f-5ff8-0cea-1bdfdb74bc9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.636883] env[61987]: DEBUG nova.network.neutron [-] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.836164] env[61987]: DEBUG nova.network.neutron [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 782.990388] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061599, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.21161} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.990725] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 782.991563] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b9707a8-401f-4284-976a-d1faa8dca918 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.015381] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 93a27dea-f28a-4cbf-ac37-ff41763ae18d/93a27dea-f28a-4cbf-ac37-ff41763ae18d.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 783.018309] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f547cb1a-7b92-4c99-9480-67a738f90a96 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.040825] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 783.040825] env[61987]: value = "task-1061600" [ 783.040825] env[61987]: _type = "Task" [ 783.040825] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.049623] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061600, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.103006] env[61987]: DEBUG nova.network.neutron [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Updating instance_info_cache with network_info: [{"id": "b9f8f891-82e4-440d-bf4d-5b3c70c910e6", "address": "fa:16:3e:0e:8f:0f", "network": {"id": "90535a3b-74ac-4bdc-a5ea-78272b01cab6", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1880668508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d8a522ba9bf422db98ca69c8d92b373", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9f8f891-82", "ovs_interfaceid": "b9f8f891-82e4-440d-bf4d-5b3c70c910e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.128208] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c7ad67-ac9f-5ff8-0cea-1bdfdb74bc9f, 'name': SearchDatastore_Task, 'duration_secs': 0.00922} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.128537] env[61987]: DEBUG oslo_concurrency.lockutils [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.128739] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 30c4fff0-c77e-4c11-aced-a040003a6b10/30c4fff0-c77e-4c11-aced-a040003a6b10.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 783.128993] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-21910f3f-8ed3-45aa-a07a-5514f596b33c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.131338] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ad40937f-d5a9-4708-8b81-06087d38d765" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.137262] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 783.137262] env[61987]: value = "task-1061601" [ 783.137262] env[61987]: _type = "Task" [ 783.137262] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.141059] env[61987]: INFO nova.compute.manager [-] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Took 1.51 seconds to deallocate network for instance. [ 783.156101] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061601, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.548691] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3805113c-b852-4bba-96bd-569a4ae9f912 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.559985] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061600, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.561068] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b780909-8ec8-43bf-a27f-028485ed04b3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.597975] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-039522c1-3288-4be4-bdf1-e38bff592685 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.607733] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Releasing lock "refresh_cache-10354ba0-5a64-4174-9c61-152d54a5dfcc" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.607882] env[61987]: DEBUG nova.compute.manager [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Instance network_info: |[{"id": "b9f8f891-82e4-440d-bf4d-5b3c70c910e6", "address": "fa:16:3e:0e:8f:0f", "network": {"id": "90535a3b-74ac-4bdc-a5ea-78272b01cab6", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1880668508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d8a522ba9bf422db98ca69c8d92b373", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9f8f891-82", "ovs_interfaceid": "b9f8f891-82e4-440d-bf4d-5b3c70c910e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 783.608637] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:8f:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60567ee6-01d0-4b16-9c7a-4a896827d6eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b9f8f891-82e4-440d-bf4d-5b3c70c910e6', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 783.618821] env[61987]: DEBUG oslo.service.loopingcall [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 783.620196] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-738b9e11-33b6-424f-a67d-8db2f7f98d38 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.625590] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 783.625881] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e6577916-67fd-4ff9-8e67-4b68ad40a112 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.656786] env[61987]: DEBUG oslo_concurrency.lockutils [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.658123] env[61987]: DEBUG nova.compute.provider_tree [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 783.664029] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 783.664029] env[61987]: value = "task-1061602" [ 783.664029] env[61987]: _type = "Task" [ 783.664029] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.668310] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061601, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.510783} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.669249] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 30c4fff0-c77e-4c11-aced-a040003a6b10/30c4fff0-c77e-4c11-aced-a040003a6b10.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 783.669481] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 783.669747] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-df9cfba1-f616-4a77-94f9-421c076e0e49 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.676083] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061602, 'name': CreateVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.681839] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 783.681839] env[61987]: value = "task-1061603" [ 783.681839] env[61987]: _type = "Task" [ 783.681839] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.690273] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061603, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.885401] env[61987]: DEBUG nova.compute.manager [req-2056e299-833a-497b-97f1-e0fe68d632ba req-d4fed7f4-1382-446d-9c9f-7836bbee71fc service nova] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Received event network-vif-deleted-92111125-1a9f-426b-85f2-0450aa02f691 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 784.029283] env[61987]: DEBUG nova.compute.manager [req-5dd799ae-0204-40da-a7a1-a66eddd5a11a req-def52518-b3c7-4530-bca8-c5862a6592da service nova] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Received event network-changed-b9f8f891-82e4-440d-bf4d-5b3c70c910e6 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 784.029833] env[61987]: DEBUG nova.compute.manager [req-5dd799ae-0204-40da-a7a1-a66eddd5a11a req-def52518-b3c7-4530-bca8-c5862a6592da service nova] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Refreshing instance network info cache due to event network-changed-b9f8f891-82e4-440d-bf4d-5b3c70c910e6. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 784.030109] env[61987]: DEBUG oslo_concurrency.lockutils [req-5dd799ae-0204-40da-a7a1-a66eddd5a11a req-def52518-b3c7-4530-bca8-c5862a6592da service nova] Acquiring lock "refresh_cache-10354ba0-5a64-4174-9c61-152d54a5dfcc" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.030276] env[61987]: DEBUG oslo_concurrency.lockutils [req-5dd799ae-0204-40da-a7a1-a66eddd5a11a req-def52518-b3c7-4530-bca8-c5862a6592da service nova] Acquired lock "refresh_cache-10354ba0-5a64-4174-9c61-152d54a5dfcc" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.030471] env[61987]: DEBUG nova.network.neutron [req-5dd799ae-0204-40da-a7a1-a66eddd5a11a req-def52518-b3c7-4530-bca8-c5862a6592da service nova] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Refreshing network info cache for port b9f8f891-82e4-440d-bf4d-5b3c70c910e6 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 784.057876] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061600, 'name': ReconfigVM_Task, 'duration_secs': 0.589348} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.057876] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 93a27dea-f28a-4cbf-ac37-ff41763ae18d/93a27dea-f28a-4cbf-ac37-ff41763ae18d.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 784.057876] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-98da328d-1e55-45dd-947e-23603ed31df2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.061976] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 784.061976] env[61987]: value = "task-1061604" [ 784.061976] env[61987]: _type = "Task" [ 784.061976] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.075261] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061604, 'name': Rename_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.163410] env[61987]: DEBUG nova.scheduler.client.report [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 784.182920] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061602, 'name': CreateVM_Task, 'duration_secs': 0.282577} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.187737] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 784.189592] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.189592] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.189724] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 784.190485] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-429d325a-b1c3-405b-9089-559037b0c81d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.198020] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061603, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079294} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.199204] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 784.199564] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for the task: (returnval){ [ 784.199564] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]526926b7-0964-45cf-da4e-42367d21860c" [ 784.199564] env[61987]: _type = "Task" [ 784.199564] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.200272] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c791f1e-3b00-4c99-b52f-85e8a1819f98 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.231542] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 30c4fff0-c77e-4c11-aced-a040003a6b10/30c4fff0-c77e-4c11-aced-a040003a6b10.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 784.236261] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-023ef5f9-c2af-4ffd-84bf-3a858bc84898 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.250118] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]526926b7-0964-45cf-da4e-42367d21860c, 'name': SearchDatastore_Task, 'duration_secs': 0.013887} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.250719] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "ad40937f-d5a9-4708-8b81-06087d38d765" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.250947] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ad40937f-d5a9-4708-8b81-06087d38d765" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.251193] env[61987]: INFO nova.compute.manager [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Attaching volume 26cf07f3-1a6a-47ba-bd88-c91cccac13bc to /dev/sdb [ 784.252847] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.253090] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 784.253316] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.253566] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.253644] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 784.255097] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ad08fa32-fc4e-4c03-a073-29cacf7e794d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.262137] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 784.262137] env[61987]: value = "task-1061605" [ 784.262137] env[61987]: _type = "Task" [ 784.262137] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.265734] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 784.265923] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 784.266920] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b567f04-d382-4583-9d27-310a6f0b156a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.272380] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061605, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.275237] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for the task: (returnval){ [ 784.275237] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]528f1ec5-9029-9bb4-fba2-1f78fe0ad89d" [ 784.275237] env[61987]: _type = "Task" [ 784.275237] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.282824] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]528f1ec5-9029-9bb4-fba2-1f78fe0ad89d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.284803] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5aaf2ad-03dc-4b2e-9a55-0c954e1fddc1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.290448] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80b6bebc-a2ad-4dd8-8bcb-8322648b69df {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.306336] env[61987]: DEBUG nova.virt.block_device [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Updating existing volume attachment record: 0a912651-7f03-4b36-9f13-92f460c74dbc {{(pid=61987) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 784.571832] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061604, 'name': Rename_Task, 'duration_secs': 0.149987} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.572346] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 784.572716] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e95890ff-54df-474a-9cf7-648435611274 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.579244] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 784.579244] env[61987]: value = "task-1061607" [ 784.579244] env[61987]: _type = "Task" [ 784.579244] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.587220] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061607, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.679911] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.674s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.679911] env[61987]: DEBUG nova.compute.manager [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 784.682096] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.157s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.682349] env[61987]: DEBUG nova.objects.instance [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lazy-loading 'resources' on Instance uuid 325183b9-e35b-4f9c-a271-d0fdb05288bb {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 784.771705] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061605, 'name': ReconfigVM_Task, 'duration_secs': 0.262675} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.772074] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 30c4fff0-c77e-4c11-aced-a040003a6b10/30c4fff0-c77e-4c11-aced-a040003a6b10.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 784.772681] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-727f0630-4e0f-4d4c-8f3a-0dd3257436cb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.779604] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 784.779604] env[61987]: value = "task-1061610" [ 784.779604] env[61987]: _type = "Task" [ 784.779604] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.786423] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]528f1ec5-9029-9bb4-fba2-1f78fe0ad89d, 'name': SearchDatastore_Task, 'duration_secs': 0.008744} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.787953] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f349c5a-4081-412a-b0a5-3966ca9a220e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.793817] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061610, 'name': Rename_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.797906] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for the task: (returnval){ [ 784.797906] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]521e0503-239d-da1a-fd37-121edb2a235d" [ 784.797906] env[61987]: _type = "Task" [ 784.797906] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.804210] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]521e0503-239d-da1a-fd37-121edb2a235d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.814826] env[61987]: DEBUG nova.network.neutron [req-5dd799ae-0204-40da-a7a1-a66eddd5a11a req-def52518-b3c7-4530-bca8-c5862a6592da service nova] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Updated VIF entry in instance network info cache for port b9f8f891-82e4-440d-bf4d-5b3c70c910e6. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 784.815191] env[61987]: DEBUG nova.network.neutron [req-5dd799ae-0204-40da-a7a1-a66eddd5a11a req-def52518-b3c7-4530-bca8-c5862a6592da service nova] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Updating instance_info_cache with network_info: [{"id": "b9f8f891-82e4-440d-bf4d-5b3c70c910e6", "address": "fa:16:3e:0e:8f:0f", "network": {"id": "90535a3b-74ac-4bdc-a5ea-78272b01cab6", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1880668508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d8a522ba9bf422db98ca69c8d92b373", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9f8f891-82", "ovs_interfaceid": "b9f8f891-82e4-440d-bf4d-5b3c70c910e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.093437] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061607, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.192992] env[61987]: DEBUG nova.compute.utils [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 785.194727] env[61987]: DEBUG nova.compute.manager [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 785.196146] env[61987]: DEBUG nova.network.neutron [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 785.266480] env[61987]: DEBUG nova.policy [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e16f557dfdb04406b3fc772b891f05a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '22e62741abfb4471bbc7e11ae7d407ca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 785.292765] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061610, 'name': Rename_Task, 'duration_secs': 0.142687} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.293166] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 785.293385] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a58f2b06-86d0-43a5-8ef0-a36785fc5525 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.306019] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 785.306019] env[61987]: value = "task-1061611" [ 785.306019] env[61987]: _type = "Task" [ 785.306019] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.310692] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]521e0503-239d-da1a-fd37-121edb2a235d, 'name': SearchDatastore_Task, 'duration_secs': 0.008473} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.313545] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.313848] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 10354ba0-5a64-4174-9c61-152d54a5dfcc/10354ba0-5a64-4174-9c61-152d54a5dfcc.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 785.314305] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d6eab94f-ea8e-4cc0-b183-0c0173435ab3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.319709] env[61987]: DEBUG oslo_concurrency.lockutils [req-5dd799ae-0204-40da-a7a1-a66eddd5a11a req-def52518-b3c7-4530-bca8-c5862a6592da service nova] Releasing lock "refresh_cache-10354ba0-5a64-4174-9c61-152d54a5dfcc" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.320098] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061611, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.324209] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for the task: (returnval){ [ 785.324209] env[61987]: value = "task-1061612" [ 785.324209] env[61987]: _type = "Task" [ 785.324209] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.341428] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061612, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.589441] env[61987]: DEBUG nova.network.neutron [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Successfully created port: c9f4e2e0-50b8-4e72-a06d-8c2729b200ed {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 785.597202] env[61987]: DEBUG oslo_vmware.api [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061607, 'name': PowerOnVM_Task, 'duration_secs': 0.519752} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.597727] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 785.597948] env[61987]: INFO nova.compute.manager [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Took 9.84 seconds to spawn the instance on the hypervisor. [ 785.598157] env[61987]: DEBUG nova.compute.manager [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 785.598979] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7264a036-e9d4-4e7a-8116-853edf86c0a5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.636971] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c7be59-696a-410b-8715-c6f5e059da7a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.647032] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f5ad3b-289d-473f-b5bd-53770856f9d2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.681681] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10995ae2-c32a-4acc-bc90-f653bc3c9375 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.689046] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db24c352-3de9-4319-b34d-056909ad4e4d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.706446] env[61987]: DEBUG nova.compute.manager [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 785.709485] env[61987]: DEBUG nova.compute.provider_tree [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 785.813959] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061611, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.833478] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061612, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.489255} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.833763] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 10354ba0-5a64-4174-9c61-152d54a5dfcc/10354ba0-5a64-4174-9c61-152d54a5dfcc.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 785.834137] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 785.834297] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-de58205f-73bd-4f5f-ae13-b3f2acfe2062 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.841510] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for the task: (returnval){ [ 785.841510] env[61987]: value = "task-1061613" [ 785.841510] env[61987]: _type = "Task" [ 785.841510] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.849657] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061613, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.122026] env[61987]: INFO nova.compute.manager [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Took 45.49 seconds to build instance. [ 786.216280] env[61987]: DEBUG nova.scheduler.client.report [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 786.317610] env[61987]: DEBUG oslo_vmware.api [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061611, 'name': PowerOnVM_Task, 'duration_secs': 0.676792} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.317773] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 786.320379] env[61987]: INFO nova.compute.manager [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Took 7.95 seconds to spawn the instance on the hypervisor. [ 786.320379] env[61987]: DEBUG nova.compute.manager [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 786.320379] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c35ffc-d623-4cd9-b51f-bfcd90710ca9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.350480] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061613, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.10886} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.351562] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 786.351636] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aeb6fb0-d59c-4d17-8ae9-95f46f045c0e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.374582] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 10354ba0-5a64-4174-9c61-152d54a5dfcc/10354ba0-5a64-4174-9c61-152d54a5dfcc.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 786.375666] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6c6960d-7534-4003-b797-029b7a7d58c0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.396193] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for the task: (returnval){ [ 786.396193] env[61987]: value = "task-1061614" [ 786.396193] env[61987]: _type = "Task" [ 786.396193] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.407665] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061614, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.628606] env[61987]: DEBUG oslo_concurrency.lockutils [None req-252dfadb-e12e-4222-ac3d-10b82b010a11 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "93a27dea-f28a-4cbf-ac37-ff41763ae18d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 74.007s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.649636] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "93a27dea-f28a-4cbf-ac37-ff41763ae18d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.650663] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "93a27dea-f28a-4cbf-ac37-ff41763ae18d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.650663] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "93a27dea-f28a-4cbf-ac37-ff41763ae18d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.650663] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "93a27dea-f28a-4cbf-ac37-ff41763ae18d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.650663] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "93a27dea-f28a-4cbf-ac37-ff41763ae18d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.656279] env[61987]: INFO nova.compute.manager [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Terminating instance [ 786.722759] env[61987]: DEBUG nova.compute.manager [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 786.725696] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.044s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.727944] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.848s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.730114] env[61987]: INFO nova.compute.claims [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 786.756772] env[61987]: DEBUG nova.virt.hardware [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 786.757083] env[61987]: DEBUG nova.virt.hardware [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 786.758127] env[61987]: DEBUG nova.virt.hardware [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 786.758424] env[61987]: DEBUG nova.virt.hardware [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 786.758529] env[61987]: DEBUG nova.virt.hardware [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 786.758688] env[61987]: DEBUG nova.virt.hardware [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 786.758906] env[61987]: DEBUG nova.virt.hardware [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 786.759098] env[61987]: DEBUG nova.virt.hardware [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 786.759281] env[61987]: DEBUG nova.virt.hardware [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 786.759491] env[61987]: DEBUG nova.virt.hardware [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 786.759926] env[61987]: DEBUG nova.virt.hardware [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 786.761157] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4dbb275-a658-42e2-a367-4c035bdbce52 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.765589] env[61987]: INFO nova.scheduler.client.report [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Deleted allocations for instance 325183b9-e35b-4f9c-a271-d0fdb05288bb [ 786.773632] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a497a7a-0d71-46dd-b417-7f2ef54c2d08 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.838303] env[61987]: INFO nova.compute.manager [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Took 45.68 seconds to build instance. [ 786.907469] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061614, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.955088] env[61987]: INFO nova.compute.manager [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Rebuilding instance [ 787.006604] env[61987]: DEBUG nova.compute.manager [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 787.007569] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77c42a5d-e476-4460-be76-c23207077e40 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.165737] env[61987]: DEBUG nova.compute.manager [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 787.165737] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 787.165737] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-528a2e6e-2aa6-45ef-8339-eff0f5a07d32 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.178578] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 787.178866] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41e45dfd-455b-49ee-911f-7d9219357ca2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.186875] env[61987]: DEBUG oslo_vmware.api [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 787.186875] env[61987]: value = "task-1061616" [ 787.186875] env[61987]: _type = "Task" [ 787.186875] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.198187] env[61987]: DEBUG oslo_vmware.api [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061616, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.232226] env[61987]: DEBUG nova.network.neutron [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Successfully updated port: c9f4e2e0-50b8-4e72-a06d-8c2729b200ed {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 787.275939] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e89de5d-a888-420a-a1a3-44978425b7c9 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "325183b9-e35b-4f9c-a271-d0fdb05288bb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.995s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.339967] env[61987]: DEBUG oslo_concurrency.lockutils [None req-afc6b200-fe70-45e2-a211-1f5b2057dd7f tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "30c4fff0-c77e-4c11-aced-a040003a6b10" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.525s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.409122] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061614, 'name': ReconfigVM_Task, 'duration_secs': 0.906481} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.409498] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 10354ba0-5a64-4174-9c61-152d54a5dfcc/10354ba0-5a64-4174-9c61-152d54a5dfcc.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 787.410117] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3a578a9d-66db-4453-aaa6-57ef755ac24e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.416087] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for the task: (returnval){ [ 787.416087] env[61987]: value = "task-1061617" [ 787.416087] env[61987]: _type = "Task" [ 787.416087] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.426343] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061617, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.488894] env[61987]: DEBUG nova.compute.manager [req-34f83da3-b68c-4da4-8729-9b79f26289b7 req-2bf3d426-362b-4836-accf-988c1a925460 service nova] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Received event network-vif-plugged-c9f4e2e0-50b8-4e72-a06d-8c2729b200ed {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 787.489192] env[61987]: DEBUG oslo_concurrency.lockutils [req-34f83da3-b68c-4da4-8729-9b79f26289b7 req-2bf3d426-362b-4836-accf-988c1a925460 service nova] Acquiring lock "84a4a69b-04db-4546-ba89-9039e382a0c4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.489446] env[61987]: DEBUG oslo_concurrency.lockutils [req-34f83da3-b68c-4da4-8729-9b79f26289b7 req-2bf3d426-362b-4836-accf-988c1a925460 service nova] Lock "84a4a69b-04db-4546-ba89-9039e382a0c4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.489656] env[61987]: DEBUG oslo_concurrency.lockutils [req-34f83da3-b68c-4da4-8729-9b79f26289b7 req-2bf3d426-362b-4836-accf-988c1a925460 service nova] Lock "84a4a69b-04db-4546-ba89-9039e382a0c4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.489840] env[61987]: DEBUG nova.compute.manager [req-34f83da3-b68c-4da4-8729-9b79f26289b7 req-2bf3d426-362b-4836-accf-988c1a925460 service nova] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] No waiting events found dispatching network-vif-plugged-c9f4e2e0-50b8-4e72-a06d-8c2729b200ed {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 787.490022] env[61987]: WARNING nova.compute.manager [req-34f83da3-b68c-4da4-8729-9b79f26289b7 req-2bf3d426-362b-4836-accf-988c1a925460 service nova] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Received unexpected event network-vif-plugged-c9f4e2e0-50b8-4e72-a06d-8c2729b200ed for instance with vm_state building and task_state spawning. [ 787.698515] env[61987]: DEBUG oslo_vmware.api [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061616, 'name': PowerOffVM_Task, 'duration_secs': 0.220889} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.698719] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 787.698895] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 787.699645] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3590f7db-ec74-46e4-b695-781790076bab {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.745774] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "refresh_cache-84a4a69b-04db-4546-ba89-9039e382a0c4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.745935] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired lock "refresh_cache-84a4a69b-04db-4546-ba89-9039e382a0c4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.746158] env[61987]: DEBUG nova.network.neutron [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 787.759548] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 787.759797] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 787.759992] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Deleting the datastore file [datastore2] 93a27dea-f28a-4cbf-ac37-ff41763ae18d {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 787.760269] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a8430448-63a3-406c-9a6d-e3bff003c336 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.768767] env[61987]: DEBUG oslo_vmware.api [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 787.768767] env[61987]: value = "task-1061619" [ 787.768767] env[61987]: _type = "Task" [ 787.768767] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.783030] env[61987]: DEBUG oslo_vmware.api [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061619, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.929328] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061617, 'name': Rename_Task, 'duration_secs': 0.145903} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.932229] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 787.932758] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-acd53848-1f33-421d-a39a-15066c41ccfc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.941233] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for the task: (returnval){ [ 787.941233] env[61987]: value = "task-1061620" [ 787.941233] env[61987]: _type = "Task" [ 787.941233] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.957794] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061620, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.029754] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 788.031784] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d0ad6f52-4c84-4615-8e37-62c5502441ab {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.038516] env[61987]: DEBUG oslo_vmware.api [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Waiting for the task: (returnval){ [ 788.038516] env[61987]: value = "task-1061621" [ 788.038516] env[61987]: _type = "Task" [ 788.038516] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.051072] env[61987]: DEBUG oslo_vmware.api [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Task: {'id': task-1061621, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.259295] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed3d7c8-f164-4d77-8230-d7c79fc34c19 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.272766] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1846e733-95b1-4372-a8c5-879d926693bf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.283815] env[61987]: DEBUG oslo_vmware.api [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061619, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150166} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.308350] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 788.308350] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 788.308350] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 788.308652] env[61987]: INFO nova.compute.manager [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Took 1.15 seconds to destroy the instance on the hypervisor. [ 788.308739] env[61987]: DEBUG oslo.service.loopingcall [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 788.309192] env[61987]: DEBUG nova.compute.manager [-] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 788.309296] env[61987]: DEBUG nova.network.neutron [-] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 788.311567] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ea51bc4-48be-43d0-861f-2333bc264627 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.315687] env[61987]: DEBUG nova.network.neutron [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 788.321064] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f865e767-e6a3-4361-a0dc-afdb73f94a48 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.336704] env[61987]: DEBUG nova.compute.provider_tree [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.414998] env[61987]: DEBUG nova.compute.manager [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 788.415939] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8fdcb80-8e83-4f45-a3c5-2ef315938f52 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.451719] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061620, 'name': PowerOnVM_Task} progress is 76%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.551178] env[61987]: DEBUG oslo_vmware.api [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Task: {'id': task-1061621, 'name': PowerOffVM_Task, 'duration_secs': 0.186024} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.554236] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 788.555348] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 788.556027] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-23da0231-bd81-42ec-86d7-8496a3dfc8af {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.564059] env[61987]: DEBUG oslo_vmware.api [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Waiting for the task: (returnval){ [ 788.564059] env[61987]: value = "task-1061622" [ 788.564059] env[61987]: _type = "Task" [ 788.564059] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.570865] env[61987]: DEBUG oslo_vmware.api [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Task: {'id': task-1061622, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.658639] env[61987]: DEBUG nova.network.neutron [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Updating instance_info_cache with network_info: [{"id": "c9f4e2e0-50b8-4e72-a06d-8c2729b200ed", "address": "fa:16:3e:74:cd:56", "network": {"id": "d5ed1e9e-9033-441e-bfcb-2ac643d7f03f", "bridge": "br-int", "label": "tempest-ServersTestJSON-8466993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e62741abfb4471bbc7e11ae7d407ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39ab9baf-90cd-4fe2-8d56-434f8210fc19", "external-id": "nsx-vlan-transportzone-713", "segmentation_id": 713, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9f4e2e0-50", "ovs_interfaceid": "c9f4e2e0-50b8-4e72-a06d-8c2729b200ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.839821] env[61987]: DEBUG nova.scheduler.client.report [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 788.856953] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Volume attach. Driver type: vmdk {{(pid=61987) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 788.857316] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234324', 'volume_id': '26cf07f3-1a6a-47ba-bd88-c91cccac13bc', 'name': 'volume-26cf07f3-1a6a-47ba-bd88-c91cccac13bc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ad40937f-d5a9-4708-8b81-06087d38d765', 'attached_at': '', 'detached_at': '', 'volume_id': '26cf07f3-1a6a-47ba-bd88-c91cccac13bc', 'serial': '26cf07f3-1a6a-47ba-bd88-c91cccac13bc'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 788.858688] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c1f82f4-0d12-47c5-83ee-000da864d508 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.879347] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f239b4db-cbac-4eb3-9c87-e10f0a1952ca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.909175] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] volume-26cf07f3-1a6a-47ba-bd88-c91cccac13bc/volume-26cf07f3-1a6a-47ba-bd88-c91cccac13bc.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 788.911117] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53bf437c-5d0e-48e2-b8f0-a936d64d15e7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.928143] env[61987]: INFO nova.compute.manager [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] instance snapshotting [ 788.934037] env[61987]: DEBUG oslo_vmware.api [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 788.934037] env[61987]: value = "task-1061623" [ 788.934037] env[61987]: _type = "Task" [ 788.934037] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.934037] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33a545dd-23e6-4a0e-ac94-562aa4a4ab64 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.964547] env[61987]: DEBUG oslo_vmware.api [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061623, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.966565] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf7cce11-dcae-457a-839f-75ea62077f90 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.974277] env[61987]: DEBUG oslo_vmware.api [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061620, 'name': PowerOnVM_Task, 'duration_secs': 0.752802} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.974277] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 788.974277] env[61987]: INFO nova.compute.manager [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Took 8.05 seconds to spawn the instance on the hypervisor. [ 788.974277] env[61987]: DEBUG nova.compute.manager [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 788.975827] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80c0f62d-2756-4e67-b45b-41651e04a79d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.073200] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] VM already powered off {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 789.073459] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Volume detach. Driver type: vmdk {{(pid=61987) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 789.073664] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234229', 'volume_id': 'a8d52043-effe-4e4a-9a52-3ed5b5050a94', 'name': 'volume-a8d52043-effe-4e4a-9a52-3ed5b5050a94', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '95641d0b-970c-4b94-8568-5c46d0808345', 'attached_at': '', 'detached_at': '', 'volume_id': 'a8d52043-effe-4e4a-9a52-3ed5b5050a94', 'serial': 'a8d52043-effe-4e4a-9a52-3ed5b5050a94'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 789.074576] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-528f2bc6-6975-48ec-a96a-be275ff6861b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.097510] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31942031-1d30-40cb-8bc2-ea5f13948de8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.103909] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5d230f1-757d-4d03-ab0b-63329fe9e36e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.124396] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f48a0c07-9304-431c-bc25-0b53cfc5ee38 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.130075] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Acquiring lock "6ed3f0e9-3c39-4406-9863-931a117961a8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.130316] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Lock "6ed3f0e9-3c39-4406-9863-931a117961a8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.145637] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] The volume has not been displaced from its original location: [datastore2] volume-a8d52043-effe-4e4a-9a52-3ed5b5050a94/volume-a8d52043-effe-4e4a-9a52-3ed5b5050a94.vmdk. No consolidation needed. {{(pid=61987) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 789.152119] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Reconfiguring VM instance instance-00000033 to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 789.153103] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd521216-bc5a-44cf-919c-9f54be440329 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.167191] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Releasing lock "refresh_cache-84a4a69b-04db-4546-ba89-9039e382a0c4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.167519] env[61987]: DEBUG nova.compute.manager [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Instance network_info: |[{"id": "c9f4e2e0-50b8-4e72-a06d-8c2729b200ed", "address": "fa:16:3e:74:cd:56", "network": {"id": "d5ed1e9e-9033-441e-bfcb-2ac643d7f03f", "bridge": "br-int", "label": "tempest-ServersTestJSON-8466993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e62741abfb4471bbc7e11ae7d407ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39ab9baf-90cd-4fe2-8d56-434f8210fc19", "external-id": "nsx-vlan-transportzone-713", "segmentation_id": 713, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9f4e2e0-50", "ovs_interfaceid": "c9f4e2e0-50b8-4e72-a06d-8c2729b200ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 789.167951] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:74:cd:56', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '39ab9baf-90cd-4fe2-8d56-434f8210fc19', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c9f4e2e0-50b8-4e72-a06d-8c2729b200ed', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 789.175534] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Creating folder: Project (22e62741abfb4471bbc7e11ae7d407ca). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 789.176624] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-30ebca28-1564-42b1-ac8a-ef5f50bb2ebc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.181642] env[61987]: DEBUG oslo_vmware.api [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Waiting for the task: (returnval){ [ 789.181642] env[61987]: value = "task-1061624" [ 789.181642] env[61987]: _type = "Task" [ 789.181642] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.186325] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Created folder: Project (22e62741abfb4471bbc7e11ae7d407ca) in parent group-v234219. [ 789.186454] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Creating folder: Instances. Parent ref: group-v234325. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 789.186980] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b235354e-d735-4748-877a-7022d86484fc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.191389] env[61987]: DEBUG oslo_vmware.api [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Task: {'id': task-1061624, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.201039] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Created folder: Instances in parent group-v234325. [ 789.201347] env[61987]: DEBUG oslo.service.loopingcall [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 789.201559] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 789.201796] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bad185fc-82b2-4d73-9af4-bcdac4308cc1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.221653] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 789.221653] env[61987]: value = "task-1061627" [ 789.221653] env[61987]: _type = "Task" [ 789.221653] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.235509] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061627, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.268911] env[61987]: DEBUG nova.network.neutron [-] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.347846] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.620s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.348531] env[61987]: DEBUG nova.compute.manager [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 789.352183] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 29.922s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.352711] env[61987]: DEBUG nova.objects.instance [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61987) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 789.448560] env[61987]: DEBUG oslo_vmware.api [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061623, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.484689] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Creating Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 789.485027] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b4aed0f9-0a12-4483-a8c3-0602795639cd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.502307] env[61987]: DEBUG oslo_vmware.api [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 789.502307] env[61987]: value = "task-1061628" [ 789.502307] env[61987]: _type = "Task" [ 789.502307] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.507040] env[61987]: INFO nova.compute.manager [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Took 44.78 seconds to build instance. [ 789.513270] env[61987]: DEBUG oslo_vmware.api [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061628, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.547208] env[61987]: DEBUG nova.compute.manager [req-8843485f-e519-4af2-9627-cdd2706355c6 req-0b89ff78-2507-4c69-aa5a-4f2ac5415170 service nova] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Received event network-changed-c9f4e2e0-50b8-4e72-a06d-8c2729b200ed {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 789.547454] env[61987]: DEBUG nova.compute.manager [req-8843485f-e519-4af2-9627-cdd2706355c6 req-0b89ff78-2507-4c69-aa5a-4f2ac5415170 service nova] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Refreshing instance network info cache due to event network-changed-c9f4e2e0-50b8-4e72-a06d-8c2729b200ed. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 789.547699] env[61987]: DEBUG oslo_concurrency.lockutils [req-8843485f-e519-4af2-9627-cdd2706355c6 req-0b89ff78-2507-4c69-aa5a-4f2ac5415170 service nova] Acquiring lock "refresh_cache-84a4a69b-04db-4546-ba89-9039e382a0c4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.547874] env[61987]: DEBUG oslo_concurrency.lockutils [req-8843485f-e519-4af2-9627-cdd2706355c6 req-0b89ff78-2507-4c69-aa5a-4f2ac5415170 service nova] Acquired lock "refresh_cache-84a4a69b-04db-4546-ba89-9039e382a0c4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.548096] env[61987]: DEBUG nova.network.neutron [req-8843485f-e519-4af2-9627-cdd2706355c6 req-0b89ff78-2507-4c69-aa5a-4f2ac5415170 service nova] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Refreshing network info cache for port c9f4e2e0-50b8-4e72-a06d-8c2729b200ed {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 789.633033] env[61987]: DEBUG nova.compute.manager [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 789.692066] env[61987]: DEBUG oslo_vmware.api [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Task: {'id': task-1061624, 'name': ReconfigVM_Task, 'duration_secs': 0.312924} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.692243] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Reconfigured VM instance instance-00000033 to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 789.697675] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1cb6db3b-ef2d-4e0a-aadd-237493f2d1c1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.712385] env[61987]: DEBUG oslo_vmware.api [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Waiting for the task: (returnval){ [ 789.712385] env[61987]: value = "task-1061629" [ 789.712385] env[61987]: _type = "Task" [ 789.712385] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.719855] env[61987]: DEBUG oslo_vmware.api [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Task: {'id': task-1061629, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.729322] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061627, 'name': CreateVM_Task, 'duration_secs': 0.391841} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.729487] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 789.730182] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.730366] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.730707] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 789.730955] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67a72785-61a0-4ae9-9910-121a70fbaf83 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.735570] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 789.735570] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]528109b1-15f5-94ac-7523-9cb7ce3d0675" [ 789.735570] env[61987]: _type = "Task" [ 789.735570] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.743512] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]528109b1-15f5-94ac-7523-9cb7ce3d0675, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.771303] env[61987]: INFO nova.compute.manager [-] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Took 1.46 seconds to deallocate network for instance. [ 789.861039] env[61987]: DEBUG nova.compute.utils [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 789.867020] env[61987]: DEBUG nova.compute.manager [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 789.867241] env[61987]: DEBUG nova.network.neutron [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 789.915025] env[61987]: DEBUG nova.policy [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a3425a6d5e36481bb702ddf789ea9eed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c3d31c138df453486a2b1a6b70a84d4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 789.947639] env[61987]: DEBUG oslo_vmware.api [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061623, 'name': ReconfigVM_Task, 'duration_secs': 0.575607} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.948662] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Reconfigured VM instance instance-00000031 to attach disk [datastore1] volume-26cf07f3-1a6a-47ba-bd88-c91cccac13bc/volume-26cf07f3-1a6a-47ba-bd88-c91cccac13bc.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 789.955380] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5d5250a7-1ddf-4456-9fb4-0dba96ce49fd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.972629] env[61987]: DEBUG oslo_vmware.api [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 789.972629] env[61987]: value = "task-1061630" [ 789.972629] env[61987]: _type = "Task" [ 789.972629] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.981848] env[61987]: DEBUG oslo_vmware.api [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061630, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.011399] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88c2f9ad-c105-496e-8d70-0f0aca34c892 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "10354ba0-5a64-4174-9c61-152d54a5dfcc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.223s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.012602] env[61987]: DEBUG oslo_vmware.api [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061628, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.153266] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.229756] env[61987]: DEBUG oslo_vmware.api [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Task: {'id': task-1061629, 'name': ReconfigVM_Task, 'duration_secs': 0.114553} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.230415] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234229', 'volume_id': 'a8d52043-effe-4e4a-9a52-3ed5b5050a94', 'name': 'volume-a8d52043-effe-4e4a-9a52-3ed5b5050a94', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '95641d0b-970c-4b94-8568-5c46d0808345', 'attached_at': '', 'detached_at': '', 'volume_id': 'a8d52043-effe-4e4a-9a52-3ed5b5050a94', 'serial': 'a8d52043-effe-4e4a-9a52-3ed5b5050a94'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 790.230765] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 790.231878] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dbe0fc5-f4ed-4607-a3ce-ac892129ccac {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.245287] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 790.245409] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-62e39fe3-74ed-4466-8517-0882937ec62f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.251227] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]528109b1-15f5-94ac-7523-9cb7ce3d0675, 'name': SearchDatastore_Task, 'duration_secs': 0.00957} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.253919] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.254218] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 790.254781] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.254781] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.254912] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 790.255486] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-41a32b97-eb7f-4f0f-a9cc-c0e4f26fddbd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.264180] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 790.264421] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 790.265224] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0a15c8b-e355-4717-bba4-fd2a248c4827 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.273273] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 790.273273] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5223a8da-ff88-edd8-1d4c-64ae5da1cf8d" [ 790.273273] env[61987]: _type = "Task" [ 790.273273] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.280034] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.286277] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5223a8da-ff88-edd8-1d4c-64ae5da1cf8d, 'name': SearchDatastore_Task, 'duration_secs': 0.008718} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.287161] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9541cd98-e0cb-40e9-9e19-6bb33211cd6d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.292700] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 790.292700] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52e43cc9-6786-78c2-985a-0371478460a9" [ 790.292700] env[61987]: _type = "Task" [ 790.292700] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.301236] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52e43cc9-6786-78c2-985a-0371478460a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.329323] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 790.329323] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 790.329323] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Deleting the datastore file [datastore2] 95641d0b-970c-4b94-8568-5c46d0808345 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 790.329323] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc62d411-12ed-4068-b3c2-eac51d165891 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.336811] env[61987]: DEBUG oslo_vmware.api [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Waiting for the task: (returnval){ [ 790.336811] env[61987]: value = "task-1061632" [ 790.336811] env[61987]: _type = "Task" [ 790.336811] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.348401] env[61987]: DEBUG oslo_vmware.api [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Task: {'id': task-1061632, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.372023] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2b8ad35c-8fa3-4004-899c-9bde951315c0 tempest-ServersAdmin275Test-841494245 tempest-ServersAdmin275Test-841494245-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.020s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.373283] env[61987]: DEBUG nova.compute.manager [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 790.379280] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 29.038s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.483348] env[61987]: DEBUG oslo_vmware.api [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061630, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.516431] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "056c181c-37da-49ac-ae1f-f7f2032c579a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.516771] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "056c181c-37da-49ac-ae1f-f7f2032c579a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.518281] env[61987]: DEBUG oslo_vmware.api [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061628, 'name': CreateSnapshot_Task, 'duration_secs': 0.984542} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.518281] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Created Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 790.520622] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f9a2034-6932-47fe-87fd-d8466696bc87 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.580898] env[61987]: DEBUG nova.network.neutron [req-8843485f-e519-4af2-9627-cdd2706355c6 req-0b89ff78-2507-4c69-aa5a-4f2ac5415170 service nova] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Updated VIF entry in instance network info cache for port c9f4e2e0-50b8-4e72-a06d-8c2729b200ed. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 790.581268] env[61987]: DEBUG nova.network.neutron [req-8843485f-e519-4af2-9627-cdd2706355c6 req-0b89ff78-2507-4c69-aa5a-4f2ac5415170 service nova] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Updating instance_info_cache with network_info: [{"id": "c9f4e2e0-50b8-4e72-a06d-8c2729b200ed", "address": "fa:16:3e:74:cd:56", "network": {"id": "d5ed1e9e-9033-441e-bfcb-2ac643d7f03f", "bridge": "br-int", "label": "tempest-ServersTestJSON-8466993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e62741abfb4471bbc7e11ae7d407ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39ab9baf-90cd-4fe2-8d56-434f8210fc19", "external-id": "nsx-vlan-transportzone-713", "segmentation_id": 713, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9f4e2e0-50", "ovs_interfaceid": "c9f4e2e0-50b8-4e72-a06d-8c2729b200ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.600404] env[61987]: DEBUG nova.network.neutron [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Successfully created port: a3d8a0f6-c3dc-42d9-b4e1-903a299b601e {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 790.807208] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52e43cc9-6786-78c2-985a-0371478460a9, 'name': SearchDatastore_Task, 'duration_secs': 0.008771} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.807540] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.807835] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 84a4a69b-04db-4546-ba89-9039e382a0c4/84a4a69b-04db-4546-ba89-9039e382a0c4.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 790.808129] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2180f213-b5eb-4085-8d5b-32f35dd55738 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.816556] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 790.816556] env[61987]: value = "task-1061633" [ 790.816556] env[61987]: _type = "Task" [ 790.816556] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.824407] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061633, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.848585] env[61987]: DEBUG oslo_vmware.api [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Task: {'id': task-1061632, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.071886} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.848849] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 790.849103] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 790.849290] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 790.884095] env[61987]: INFO nova.compute.claims [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 790.925471] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Volume detach. Driver type: vmdk {{(pid=61987) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 790.925814] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-292c778c-5c0d-4e74-a43a-90205b2ffe2d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.935270] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9dad38-b72d-4702-9939-7fafc999e8f0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.967703] env[61987]: ERROR nova.compute.manager [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Failed to detach volume a8d52043-effe-4e4a-9a52-3ed5b5050a94 from /dev/sda: nova.exception.InstanceNotFound: Instance 95641d0b-970c-4b94-8568-5c46d0808345 could not be found. [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Traceback (most recent call last): [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/compute/manager.py", line 4184, in _do_rebuild_instance [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] self.driver.rebuild(**kwargs) [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/virt/driver.py", line 497, in rebuild [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] raise NotImplementedError() [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] NotImplementedError [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] During handling of the above exception, another exception occurred: [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Traceback (most recent call last): [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/compute/manager.py", line 3607, in _detach_root_volume [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] self.driver.detach_volume(context, old_connection_info, [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 555, in detach_volume [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] return self._volumeops.detach_volume(connection_info, instance) [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] self._detach_volume_vmdk(connection_info, instance) [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] stable_ref.fetch_moref(session) [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] raise exception.InstanceNotFound(instance_id=self._uuid) [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] nova.exception.InstanceNotFound: Instance 95641d0b-970c-4b94-8568-5c46d0808345 could not be found. [ 790.967703] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] [ 790.983302] env[61987]: DEBUG oslo_vmware.api [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061630, 'name': ReconfigVM_Task, 'duration_secs': 0.943972} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.983631] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234324', 'volume_id': '26cf07f3-1a6a-47ba-bd88-c91cccac13bc', 'name': 'volume-26cf07f3-1a6a-47ba-bd88-c91cccac13bc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ad40937f-d5a9-4708-8b81-06087d38d765', 'attached_at': '', 'detached_at': '', 'volume_id': '26cf07f3-1a6a-47ba-bd88-c91cccac13bc', 'serial': '26cf07f3-1a6a-47ba-bd88-c91cccac13bc'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 791.019512] env[61987]: DEBUG nova.compute.manager [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 791.038745] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Creating linked-clone VM from snapshot {{(pid=61987) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 791.039447] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-cf72df2c-3fcd-43a7-8053-10ee86cce449 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.048684] env[61987]: DEBUG oslo_vmware.api [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 791.048684] env[61987]: value = "task-1061634" [ 791.048684] env[61987]: _type = "Task" [ 791.048684] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.058836] env[61987]: DEBUG oslo_vmware.api [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061634, 'name': CloneVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.086620] env[61987]: DEBUG oslo_concurrency.lockutils [req-8843485f-e519-4af2-9627-cdd2706355c6 req-0b89ff78-2507-4c69-aa5a-4f2ac5415170 service nova] Releasing lock "refresh_cache-84a4a69b-04db-4546-ba89-9039e382a0c4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.086951] env[61987]: DEBUG nova.compute.manager [req-8843485f-e519-4af2-9627-cdd2706355c6 req-0b89ff78-2507-4c69-aa5a-4f2ac5415170 service nova] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Received event network-vif-deleted-7eb7297c-3eaf-4825-b148-c0e6a54ec595 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 791.171653] env[61987]: DEBUG nova.compute.utils [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Build of instance 95641d0b-970c-4b94-8568-5c46d0808345 aborted: Failed to rebuild volume backed instance. {{(pid=61987) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 791.174647] env[61987]: ERROR nova.compute.manager [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 95641d0b-970c-4b94-8568-5c46d0808345 aborted: Failed to rebuild volume backed instance. [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Traceback (most recent call last): [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/compute/manager.py", line 4184, in _do_rebuild_instance [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] self.driver.rebuild(**kwargs) [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/virt/driver.py", line 497, in rebuild [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] raise NotImplementedError() [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] NotImplementedError [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] During handling of the above exception, another exception occurred: [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Traceback (most recent call last): [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/compute/manager.py", line 3642, in _rebuild_volume_backed_instance [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] self._detach_root_volume(context, instance, root_bdm) [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/compute/manager.py", line 3621, in _detach_root_volume [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] with excutils.save_and_reraise_exception(): [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] self.force_reraise() [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] raise self.value [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/compute/manager.py", line 3607, in _detach_root_volume [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] self.driver.detach_volume(context, old_connection_info, [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 555, in detach_volume [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] return self._volumeops.detach_volume(connection_info, instance) [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] self._detach_volume_vmdk(connection_info, instance) [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] stable_ref.fetch_moref(session) [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] raise exception.InstanceNotFound(instance_id=self._uuid) [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] nova.exception.InstanceNotFound: Instance 95641d0b-970c-4b94-8568-5c46d0808345 could not be found. [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] During handling of the above exception, another exception occurred: [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Traceback (most recent call last): [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/compute/manager.py", line 11338, in _error_out_instance_on_exception [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] yield [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/compute/manager.py", line 3910, in rebuild_instance [ 791.174647] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] self._do_rebuild_instance_with_claim( [ 791.175978] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/compute/manager.py", line 3996, in _do_rebuild_instance_with_claim [ 791.175978] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] self._do_rebuild_instance( [ 791.175978] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/compute/manager.py", line 4188, in _do_rebuild_instance [ 791.175978] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] self._rebuild_default_impl(**kwargs) [ 791.175978] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/compute/manager.py", line 3765, in _rebuild_default_impl [ 791.175978] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] self._rebuild_volume_backed_instance( [ 791.175978] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] File "/opt/stack/nova/nova/compute/manager.py", line 3657, in _rebuild_volume_backed_instance [ 791.175978] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] raise exception.BuildAbortException( [ 791.175978] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] nova.exception.BuildAbortException: Build of instance 95641d0b-970c-4b94-8568-5c46d0808345 aborted: Failed to rebuild volume backed instance. [ 791.175978] env[61987]: ERROR nova.compute.manager [instance: 95641d0b-970c-4b94-8568-5c46d0808345] [ 791.328373] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061633, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.398612] env[61987]: INFO nova.compute.resource_tracker [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Updating resource usage from migration c7ecbdeb-124a-4889-b4e4-2db11acba530 [ 791.401668] env[61987]: DEBUG nova.compute.manager [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 791.440214] env[61987]: DEBUG nova.virt.hardware [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 791.440542] env[61987]: DEBUG nova.virt.hardware [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 791.440742] env[61987]: DEBUG nova.virt.hardware [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 791.440965] env[61987]: DEBUG nova.virt.hardware [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 791.442464] env[61987]: DEBUG nova.virt.hardware [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 791.442464] env[61987]: DEBUG nova.virt.hardware [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 791.442464] env[61987]: DEBUG nova.virt.hardware [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 791.442464] env[61987]: DEBUG nova.virt.hardware [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 791.442464] env[61987]: DEBUG nova.virt.hardware [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 791.442464] env[61987]: DEBUG nova.virt.hardware [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 791.442892] env[61987]: DEBUG nova.virt.hardware [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 791.445057] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc1c47f1-3c47-4821-9722-ec47d13b38d6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.456674] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-162d53aa-cece-4583-88c6-7a7ca6f4ea86 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.540999] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.561208] env[61987]: DEBUG oslo_vmware.api [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061634, 'name': CloneVM_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.807314] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-536e4386-6545-4b79-a52a-8139d88ff38f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.814469] env[61987]: DEBUG nova.compute.manager [req-26f634e3-ac4e-44ef-bdd3-1d55157a6caa req-8b238048-a498-4185-9d3a-f08a306ae6f9 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Received event network-changed-39309545-c724-44c7-932e-7784c9e31fa7 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 791.814672] env[61987]: DEBUG nova.compute.manager [req-26f634e3-ac4e-44ef-bdd3-1d55157a6caa req-8b238048-a498-4185-9d3a-f08a306ae6f9 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Refreshing instance network info cache due to event network-changed-39309545-c724-44c7-932e-7784c9e31fa7. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 791.814895] env[61987]: DEBUG oslo_concurrency.lockutils [req-26f634e3-ac4e-44ef-bdd3-1d55157a6caa req-8b238048-a498-4185-9d3a-f08a306ae6f9 service nova] Acquiring lock "refresh_cache-4edd0159-6db1-41bd-a616-31a34e984059" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.815060] env[61987]: DEBUG oslo_concurrency.lockutils [req-26f634e3-ac4e-44ef-bdd3-1d55157a6caa req-8b238048-a498-4185-9d3a-f08a306ae6f9 service nova] Acquired lock "refresh_cache-4edd0159-6db1-41bd-a616-31a34e984059" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.815270] env[61987]: DEBUG nova.network.neutron [req-26f634e3-ac4e-44ef-bdd3-1d55157a6caa req-8b238048-a498-4185-9d3a-f08a306ae6f9 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Refreshing network info cache for port 39309545-c724-44c7-932e-7784c9e31fa7 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 791.824793] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e422c206-f328-4db2-b88d-e510983de5a7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.831896] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061633, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.571403} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.832519] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 84a4a69b-04db-4546-ba89-9039e382a0c4/84a4a69b-04db-4546-ba89-9039e382a0c4.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 791.832769] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 791.833249] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ccde1b44-6875-408b-b7c6-84b3eb626c5e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.865056] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02fdfe87-a123-46bd-af38-47dd0c06406a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.870934] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 791.870934] env[61987]: value = "task-1061635" [ 791.870934] env[61987]: _type = "Task" [ 791.870934] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.877471] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc6567f-ed52-44d7-91bb-e66223d55a51 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.885189] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061635, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.895355] env[61987]: DEBUG nova.compute.provider_tree [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.031108] env[61987]: DEBUG nova.objects.instance [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lazy-loading 'flavor' on Instance uuid ad40937f-d5a9-4708-8b81-06087d38d765 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 792.060856] env[61987]: DEBUG oslo_vmware.api [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061634, 'name': CloneVM_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.381939] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061635, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.252831} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.382668] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 792.383303] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01d34c4-c111-428e-92d0-8eabaf49f807 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.398578] env[61987]: DEBUG nova.scheduler.client.report [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 792.411448] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 84a4a69b-04db-4546-ba89-9039e382a0c4/84a4a69b-04db-4546-ba89-9039e382a0c4.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 792.414964] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c5c86c1-42d9-4306-bd49-3f3f520da78f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.432885] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 792.432885] env[61987]: value = "task-1061636" [ 792.432885] env[61987]: _type = "Task" [ 792.432885] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.442715] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061636, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.536445] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bba80fce-2491-464c-90ff-38349e4897b8 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ad40937f-d5a9-4708-8b81-06087d38d765" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.285s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.559573] env[61987]: DEBUG oslo_vmware.api [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061634, 'name': CloneVM_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.776025] env[61987]: DEBUG oslo_concurrency.lockutils [None req-23a529fa-a627-4c5e-93fb-d58e64ed3582 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "ad40937f-d5a9-4708-8b81-06087d38d765" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.776025] env[61987]: DEBUG oslo_concurrency.lockutils [None req-23a529fa-a627-4c5e-93fb-d58e64ed3582 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ad40937f-d5a9-4708-8b81-06087d38d765" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.776025] env[61987]: DEBUG nova.compute.manager [None req-23a529fa-a627-4c5e-93fb-d58e64ed3582 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 792.776025] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6b03a7-cc56-4514-b7b5-15cf6bfaf43e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.785633] env[61987]: DEBUG nova.compute.manager [None req-23a529fa-a627-4c5e-93fb-d58e64ed3582 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61987) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 792.785633] env[61987]: DEBUG nova.objects.instance [None req-23a529fa-a627-4c5e-93fb-d58e64ed3582 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lazy-loading 'flavor' on Instance uuid ad40937f-d5a9-4708-8b81-06087d38d765 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 792.803549] env[61987]: DEBUG nova.network.neutron [req-26f634e3-ac4e-44ef-bdd3-1d55157a6caa req-8b238048-a498-4185-9d3a-f08a306ae6f9 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Updated VIF entry in instance network info cache for port 39309545-c724-44c7-932e-7784c9e31fa7. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 792.803549] env[61987]: DEBUG nova.network.neutron [req-26f634e3-ac4e-44ef-bdd3-1d55157a6caa req-8b238048-a498-4185-9d3a-f08a306ae6f9 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Updating instance_info_cache with network_info: [{"id": "39309545-c724-44c7-932e-7784c9e31fa7", "address": "fa:16:3e:fd:80:60", "network": {"id": "90535a3b-74ac-4bdc-a5ea-78272b01cab6", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1880668508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d8a522ba9bf422db98ca69c8d92b373", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39309545-c7", "ovs_interfaceid": "39309545-c724-44c7-932e-7784c9e31fa7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.905799] env[61987]: DEBUG nova.network.neutron [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Successfully updated port: a3d8a0f6-c3dc-42d9-b4e1-903a299b601e {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 792.916142] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.537s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.916142] env[61987]: INFO nova.compute.manager [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Migrating [ 792.925152] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.476s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.926307] env[61987]: INFO nova.compute.claims [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 792.949772] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061636, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.063971] env[61987]: DEBUG oslo_vmware.api [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061634, 'name': CloneVM_Task} progress is 95%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.190784] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.303704] env[61987]: DEBUG oslo_concurrency.lockutils [req-26f634e3-ac4e-44ef-bdd3-1d55157a6caa req-8b238048-a498-4185-9d3a-f08a306ae6f9 service nova] Releasing lock "refresh_cache-4edd0159-6db1-41bd-a616-31a34e984059" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.407484] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "refresh_cache-4fd8fc17-fb13-42e7-9879-9f120fdcb330" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.407722] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquired lock "refresh_cache-4fd8fc17-fb13-42e7-9879-9f120fdcb330" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.408783] env[61987]: DEBUG nova.network.neutron [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 793.442023] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "refresh_cache-8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.442023] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired lock "refresh_cache-8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.442252] env[61987]: DEBUG nova.network.neutron [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 793.450500] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061636, 'name': ReconfigVM_Task, 'duration_secs': 0.80956} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.450772] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 84a4a69b-04db-4546-ba89-9039e382a0c4/84a4a69b-04db-4546-ba89-9039e382a0c4.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 793.451651] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6b6cac27-0b21-401e-a8ea-fc300e6964aa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.458952] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 793.458952] env[61987]: value = "task-1061637" [ 793.458952] env[61987]: _type = "Task" [ 793.458952] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.467660] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061637, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.566631] env[61987]: DEBUG oslo_vmware.api [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061634, 'name': CloneVM_Task, 'duration_secs': 2.235115} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.567046] env[61987]: INFO nova.virt.vmwareapi.vmops [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Created linked-clone VM from snapshot [ 793.568605] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-994457bc-48ae-44d2-a102-b4841e449f8c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.577737] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Uploading image 92117f13-9c42-454c-87bb-8c7e90801e79 {{(pid=61987) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 793.604405] env[61987]: DEBUG oslo_vmware.rw_handles [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 793.604405] env[61987]: value = "vm-234329" [ 793.604405] env[61987]: _type = "VirtualMachine" [ 793.604405] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 793.604587] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-11d44dd9-96a7-4575-95c3-c5d0e2d19a7c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.613070] env[61987]: DEBUG oslo_vmware.rw_handles [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lease: (returnval){ [ 793.613070] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]528f886a-e54d-f51a-81b8-5b4547d9d483" [ 793.613070] env[61987]: _type = "HttpNfcLease" [ 793.613070] env[61987]: } obtained for exporting VM: (result){ [ 793.613070] env[61987]: value = "vm-234329" [ 793.613070] env[61987]: _type = "VirtualMachine" [ 793.613070] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 793.613070] env[61987]: DEBUG oslo_vmware.api [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the lease: (returnval){ [ 793.613070] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]528f886a-e54d-f51a-81b8-5b4547d9d483" [ 793.613070] env[61987]: _type = "HttpNfcLease" [ 793.613070] env[61987]: } to be ready. {{(pid=61987) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 793.618899] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 793.618899] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]528f886a-e54d-f51a-81b8-5b4547d9d483" [ 793.618899] env[61987]: _type = "HttpNfcLease" [ 793.618899] env[61987]: } is initializing. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 793.793686] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-23a529fa-a627-4c5e-93fb-d58e64ed3582 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 793.794154] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b7356b0-0c41-4696-8332-14dcfe1957b0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.803806] env[61987]: DEBUG oslo_vmware.api [None req-23a529fa-a627-4c5e-93fb-d58e64ed3582 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 793.803806] env[61987]: value = "task-1061639" [ 793.803806] env[61987]: _type = "Task" [ 793.803806] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.816227] env[61987]: DEBUG oslo_vmware.api [None req-23a529fa-a627-4c5e-93fb-d58e64ed3582 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061639, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.969493] env[61987]: DEBUG nova.compute.manager [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Received event network-changed-39309545-c724-44c7-932e-7784c9e31fa7 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 793.969712] env[61987]: DEBUG nova.compute.manager [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Refreshing instance network info cache due to event network-changed-39309545-c724-44c7-932e-7784c9e31fa7. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 793.969943] env[61987]: DEBUG oslo_concurrency.lockutils [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] Acquiring lock "refresh_cache-4edd0159-6db1-41bd-a616-31a34e984059" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.970113] env[61987]: DEBUG oslo_concurrency.lockutils [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] Acquired lock "refresh_cache-4edd0159-6db1-41bd-a616-31a34e984059" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.970285] env[61987]: DEBUG nova.network.neutron [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Refreshing network info cache for port 39309545-c724-44c7-932e-7784c9e31fa7 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 793.985022] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061637, 'name': Rename_Task, 'duration_secs': 0.171671} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.985920] env[61987]: DEBUG nova.network.neutron [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 793.989520] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 793.991399] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ca31a9dc-1890-44a8-b845-e5ab5f587d9d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.001547] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 794.001547] env[61987]: value = "task-1061640" [ 794.001547] env[61987]: _type = "Task" [ 794.001547] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.009497] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061640, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.124286] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 794.124286] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]528f886a-e54d-f51a-81b8-5b4547d9d483" [ 794.124286] env[61987]: _type = "HttpNfcLease" [ 794.124286] env[61987]: } is ready. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 794.124600] env[61987]: DEBUG oslo_vmware.rw_handles [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 794.124600] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]528f886a-e54d-f51a-81b8-5b4547d9d483" [ 794.124600] env[61987]: _type = "HttpNfcLease" [ 794.124600] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 794.125554] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de32de2f-093f-4641-91f0-cd00fb661f2e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.137340] env[61987]: DEBUG oslo_vmware.rw_handles [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293ff7f-4a7a-78b8-22eb-53d6116ccfe1/disk-0.vmdk from lease info. {{(pid=61987) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 794.137340] env[61987]: DEBUG oslo_vmware.rw_handles [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293ff7f-4a7a-78b8-22eb-53d6116ccfe1/disk-0.vmdk for reading. {{(pid=61987) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 794.258727] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e12efc24-bf1a-449a-bc12-a3ec57687e32 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.324598] env[61987]: DEBUG oslo_vmware.api [None req-23a529fa-a627-4c5e-93fb-d58e64ed3582 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061639, 'name': PowerOffVM_Task, 'duration_secs': 0.202516} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.324894] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-23a529fa-a627-4c5e-93fb-d58e64ed3582 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 794.325108] env[61987]: DEBUG nova.compute.manager [None req-23a529fa-a627-4c5e-93fb-d58e64ed3582 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 794.325967] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f7993b2-fe55-43b6-8028-7518e6651313 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.421261] env[61987]: DEBUG nova.network.neutron [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Updating instance_info_cache with network_info: [{"id": "a3d8a0f6-c3dc-42d9-b4e1-903a299b601e", "address": "fa:16:3e:86:34:de", "network": {"id": "8dfc473a-6d57-4169-9338-49a9565bb93d", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-422412073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c3d31c138df453486a2b1a6b70a84d4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3d8a0f6-c3", "ovs_interfaceid": "a3d8a0f6-c3dc-42d9-b4e1-903a299b601e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.461609] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6374ed74-3f70-436b-a703-08e2cddd75bf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.469951] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e8b3ed-7a4e-4b82-a0e7-69b9bd2ef4aa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.511671] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8494cad4-bb7d-402d-bb8e-090333cf26be {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.516836] env[61987]: DEBUG nova.network.neutron [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Updating instance_info_cache with network_info: [{"id": "55e46e63-9076-400e-a73a-19fc4bb21bec", "address": "fa:16:3e:09:61:0c", "network": {"id": "59d8d7d6-fd71-4dec-802e-e0f9e670ff70", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1845227031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "315a2a3450024ed38e27cd7758f392f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55e46e63-90", "ovs_interfaceid": "55e46e63-9076-400e-a73a-19fc4bb21bec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.525022] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061640, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.525437] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2fe7a6e-48e2-4e23-89ee-c80906f31fe9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.545543] env[61987]: DEBUG nova.compute.provider_tree [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.549142] env[61987]: DEBUG oslo_concurrency.lockutils [None req-acb92527-6504-4fe5-bca2-a703c7bd5326 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Acquiring lock "95641d0b-970c-4b94-8568-5c46d0808345" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.549449] env[61987]: DEBUG oslo_concurrency.lockutils [None req-acb92527-6504-4fe5-bca2-a703c7bd5326 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Lock "95641d0b-970c-4b94-8568-5c46d0808345" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.549613] env[61987]: DEBUG oslo_concurrency.lockutils [None req-acb92527-6504-4fe5-bca2-a703c7bd5326 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Acquiring lock "95641d0b-970c-4b94-8568-5c46d0808345-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.549941] env[61987]: DEBUG oslo_concurrency.lockutils [None req-acb92527-6504-4fe5-bca2-a703c7bd5326 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Lock "95641d0b-970c-4b94-8568-5c46d0808345-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.550432] env[61987]: DEBUG oslo_concurrency.lockutils [None req-acb92527-6504-4fe5-bca2-a703c7bd5326 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Lock "95641d0b-970c-4b94-8568-5c46d0808345-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.555216] env[61987]: INFO nova.compute.manager [None req-acb92527-6504-4fe5-bca2-a703c7bd5326 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Terminating instance [ 794.843268] env[61987]: DEBUG oslo_concurrency.lockutils [None req-23a529fa-a627-4c5e-93fb-d58e64ed3582 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ad40937f-d5a9-4708-8b81-06087d38d765" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.068s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.880941] env[61987]: DEBUG nova.network.neutron [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Updated VIF entry in instance network info cache for port 39309545-c724-44c7-932e-7784c9e31fa7. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 794.881545] env[61987]: DEBUG nova.network.neutron [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Updating instance_info_cache with network_info: [{"id": "39309545-c724-44c7-932e-7784c9e31fa7", "address": "fa:16:3e:fd:80:60", "network": {"id": "90535a3b-74ac-4bdc-a5ea-78272b01cab6", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1880668508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d8a522ba9bf422db98ca69c8d92b373", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39309545-c7", "ovs_interfaceid": "39309545-c724-44c7-932e-7784c9e31fa7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.924860] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Releasing lock "refresh_cache-4fd8fc17-fb13-42e7-9879-9f120fdcb330" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.925331] env[61987]: DEBUG nova.compute.manager [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Instance network_info: |[{"id": "a3d8a0f6-c3dc-42d9-b4e1-903a299b601e", "address": "fa:16:3e:86:34:de", "network": {"id": "8dfc473a-6d57-4169-9338-49a9565bb93d", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-422412073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c3d31c138df453486a2b1a6b70a84d4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3d8a0f6-c3", "ovs_interfaceid": "a3d8a0f6-c3dc-42d9-b4e1-903a299b601e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 794.925975] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:34:de', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae4e3171-21cd-4094-b6cf-81bf366c75bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a3d8a0f6-c3dc-42d9-b4e1-903a299b601e', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 794.934748] env[61987]: DEBUG oslo.service.loopingcall [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.938875] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 794.939252] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-21e23cda-1609-44cd-b6b1-ea69c10f315a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.962480] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 794.962480] env[61987]: value = "task-1061641" [ 794.962480] env[61987]: _type = "Task" [ 794.962480] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.971179] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061641, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.014614] env[61987]: DEBUG oslo_vmware.api [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061640, 'name': PowerOnVM_Task, 'duration_secs': 0.649789} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.015056] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 795.015370] env[61987]: INFO nova.compute.manager [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Took 8.29 seconds to spawn the instance on the hypervisor. [ 795.015646] env[61987]: DEBUG nova.compute.manager [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 795.016472] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e23ccf-2791-44c5-b894-b16ef448bfa4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.019536] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Releasing lock "refresh_cache-8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.049442] env[61987]: DEBUG nova.scheduler.client.report [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 795.065501] env[61987]: DEBUG nova.compute.manager [None req-acb92527-6504-4fe5-bca2-a703c7bd5326 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 795.066405] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d9298a32-7b77-4845-8cf4-13acd37575c6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.075757] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8ace024-2c2a-4e98-b427-f74ec389930c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.111148] env[61987]: WARNING nova.virt.vmwareapi.driver [None req-acb92527-6504-4fe5-bca2-a703c7bd5326 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 95641d0b-970c-4b94-8568-5c46d0808345 could not be found. [ 795.111505] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-acb92527-6504-4fe5-bca2-a703c7bd5326 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 795.112367] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8d648251-33ff-430f-a809-9bb423a387ae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.121635] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ae7e839-4401-45d4-8d96-4be6fd9a5d0d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.159411] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-acb92527-6504-4fe5-bca2-a703c7bd5326 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 95641d0b-970c-4b94-8568-5c46d0808345 could not be found. [ 795.159875] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-acb92527-6504-4fe5-bca2-a703c7bd5326 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 795.160677] env[61987]: INFO nova.compute.manager [None req-acb92527-6504-4fe5-bca2-a703c7bd5326 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Took 0.09 seconds to destroy the instance on the hypervisor. [ 795.161330] env[61987]: DEBUG oslo.service.loopingcall [None req-acb92527-6504-4fe5-bca2-a703c7bd5326 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 795.161781] env[61987]: DEBUG nova.compute.manager [-] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 795.161927] env[61987]: DEBUG nova.network.neutron [-] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 795.386021] env[61987]: DEBUG oslo_concurrency.lockutils [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] Releasing lock "refresh_cache-4edd0159-6db1-41bd-a616-31a34e984059" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.386139] env[61987]: DEBUG nova.compute.manager [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Received event network-changed-b9f8f891-82e4-440d-bf4d-5b3c70c910e6 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 795.386551] env[61987]: DEBUG nova.compute.manager [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Refreshing instance network info cache due to event network-changed-b9f8f891-82e4-440d-bf4d-5b3c70c910e6. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 795.387360] env[61987]: DEBUG oslo_concurrency.lockutils [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] Acquiring lock "refresh_cache-10354ba0-5a64-4174-9c61-152d54a5dfcc" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.387519] env[61987]: DEBUG oslo_concurrency.lockutils [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] Acquired lock "refresh_cache-10354ba0-5a64-4174-9c61-152d54a5dfcc" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.387693] env[61987]: DEBUG nova.network.neutron [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Refreshing network info cache for port b9f8f891-82e4-440d-bf4d-5b3c70c910e6 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 795.478635] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061641, 'name': CreateVM_Task, 'duration_secs': 0.360819} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.479267] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 795.480263] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.480580] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.481878] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 795.481878] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a7a254f-5546-4b6a-bc37-7d09bf251246 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.488508] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 795.488508] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52c0e7ec-433a-c240-f6a0-98c53691fc6a" [ 795.488508] env[61987]: _type = "Task" [ 795.488508] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.502633] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c0e7ec-433a-c240-f6a0-98c53691fc6a, 'name': SearchDatastore_Task, 'duration_secs': 0.010521} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.503220] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.503615] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 795.504162] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.504535] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.504919] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 795.505420] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-de043a7f-397d-463f-8d8a-59137defab4c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.516218] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 795.516642] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 795.517963] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9df5a41a-8025-407d-84ca-cd922fc1799d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.543957] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 795.543957] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52f567db-6090-c821-ab54-be46173294f8" [ 795.543957] env[61987]: _type = "Task" [ 795.543957] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.547397] env[61987]: INFO nova.compute.manager [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Took 46.73 seconds to build instance. [ 795.555507] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.630s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.555507] env[61987]: DEBUG nova.compute.manager [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 795.563865] env[61987]: DEBUG oslo_concurrency.lockutils [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.835s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.564189] env[61987]: DEBUG nova.objects.instance [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Lazy-loading 'resources' on Instance uuid 1afcffda-0bd2-4cbe-8cfb-12a91bb50975 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 795.565567] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52f567db-6090-c821-ab54-be46173294f8, 'name': SearchDatastore_Task, 'duration_secs': 0.00909} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.570031] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d04554d-0ad9-4902-beb3-2f4ab8dd339c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.573805] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 795.573805] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52c15f52-f0df-6c25-e589-135afa602631" [ 795.573805] env[61987]: _type = "Task" [ 795.573805] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.583968] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c15f52-f0df-6c25-e589-135afa602631, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.959965] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquiring lock "10354ba0-5a64-4174-9c61-152d54a5dfcc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.960344] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "10354ba0-5a64-4174-9c61-152d54a5dfcc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.960598] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquiring lock "10354ba0-5a64-4174-9c61-152d54a5dfcc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.960833] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "10354ba0-5a64-4174-9c61-152d54a5dfcc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.961048] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "10354ba0-5a64-4174-9c61-152d54a5dfcc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.964606] env[61987]: INFO nova.compute.manager [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Terminating instance [ 796.054209] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7adec17b-f644-4daa-8c49-6d0c08c62c03 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "84a4a69b-04db-4546-ba89-9039e382a0c4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.979s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.069029] env[61987]: DEBUG nova.compute.utils [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 796.077227] env[61987]: DEBUG nova.compute.manager [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 796.077227] env[61987]: DEBUG nova.network.neutron [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 796.101599] env[61987]: DEBUG nova.objects.instance [None req-6a08f2fb-6443-43ca-aa8c-0a34883d3e6c tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lazy-loading 'flavor' on Instance uuid ad40937f-d5a9-4708-8b81-06087d38d765 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 796.103654] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c15f52-f0df-6c25-e589-135afa602631, 'name': SearchDatastore_Task, 'duration_secs': 0.009053} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.111021] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.111021] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 4fd8fc17-fb13-42e7-9879-9f120fdcb330/4fd8fc17-fb13-42e7-9879-9f120fdcb330.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 796.111021] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f73b71dc-082e-4da0-9d7f-f1da22b64ed1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.126961] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 796.126961] env[61987]: value = "task-1061642" [ 796.126961] env[61987]: _type = "Task" [ 796.126961] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.139558] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061642, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.176514] env[61987]: DEBUG nova.policy [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '697f9b1e3b3840369549a4f1ae3173ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd2afdc9bddba4ccab2b8fdb1f31da2b6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 796.398523] env[61987]: DEBUG nova.compute.manager [req-0e41a055-ddec-4fba-b221-0ceb8542ffdf req-19ae98b8-e3bd-4dcf-b3ee-2a62c488a482 service nova] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Received event network-changed-b9f8f891-82e4-440d-bf4d-5b3c70c910e6 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 796.398740] env[61987]: DEBUG nova.compute.manager [req-0e41a055-ddec-4fba-b221-0ceb8542ffdf req-19ae98b8-e3bd-4dcf-b3ee-2a62c488a482 service nova] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Refreshing instance network info cache due to event network-changed-b9f8f891-82e4-440d-bf4d-5b3c70c910e6. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 796.399194] env[61987]: DEBUG oslo_concurrency.lockutils [req-0e41a055-ddec-4fba-b221-0ceb8542ffdf req-19ae98b8-e3bd-4dcf-b3ee-2a62c488a482 service nova] Acquiring lock "refresh_cache-10354ba0-5a64-4174-9c61-152d54a5dfcc" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.426419] env[61987]: DEBUG nova.network.neutron [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Updated VIF entry in instance network info cache for port b9f8f891-82e4-440d-bf4d-5b3c70c910e6. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 796.426907] env[61987]: DEBUG nova.network.neutron [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Updating instance_info_cache with network_info: [{"id": "b9f8f891-82e4-440d-bf4d-5b3c70c910e6", "address": "fa:16:3e:0e:8f:0f", "network": {"id": "90535a3b-74ac-4bdc-a5ea-78272b01cab6", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1880668508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d8a522ba9bf422db98ca69c8d92b373", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9f8f891-82", "ovs_interfaceid": "b9f8f891-82e4-440d-bf4d-5b3c70c910e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.469692] env[61987]: DEBUG nova.compute.manager [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 796.469912] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 796.470952] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-269dcfbb-f639-4fed-90e2-2d39dc022a26 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.484924] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 796.485855] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f112a16-ea1a-437e-ab08-17e1c15a86d7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.493401] env[61987]: DEBUG oslo_vmware.api [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for the task: (returnval){ [ 796.493401] env[61987]: value = "task-1061643" [ 796.493401] env[61987]: _type = "Task" [ 796.493401] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.506032] env[61987]: DEBUG oslo_vmware.api [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061643, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.557435] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17437cd6-0a1b-470d-9dd3-f827904b3656 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.582033] env[61987]: DEBUG nova.compute.manager [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 796.588025] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Updating instance '8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2' progress to 0 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 796.590643] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d193005-d229-4caf-9b74-e31777aed70e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.599393] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-245a0e60-25cd-471e-bdc1-dfa60d5c4a13 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.637161] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a08f2fb-6443-43ca-aa8c-0a34883d3e6c tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "refresh_cache-ad40937f-d5a9-4708-8b81-06087d38d765" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.637161] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a08f2fb-6443-43ca-aa8c-0a34883d3e6c tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquired lock "refresh_cache-ad40937f-d5a9-4708-8b81-06087d38d765" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.637161] env[61987]: DEBUG nova.network.neutron [None req-6a08f2fb-6443-43ca-aa8c-0a34883d3e6c tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 796.637590] env[61987]: DEBUG nova.objects.instance [None req-6a08f2fb-6443-43ca-aa8c-0a34883d3e6c tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lazy-loading 'info_cache' on Instance uuid ad40937f-d5a9-4708-8b81-06087d38d765 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 796.642348] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf241a23-2450-47f0-a7b3-f58c664ecc95 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.654546] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061642, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.515056} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.657831] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9a29ccc-f1b2-448d-9ff6-e9b1a8367074 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.661819] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 4fd8fc17-fb13-42e7-9879-9f120fdcb330/4fd8fc17-fb13-42e7-9879-9f120fdcb330.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 796.661819] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 796.662264] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6263b5c8-8524-4475-b9e3-572aa3152567 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.676543] env[61987]: DEBUG nova.compute.provider_tree [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 796.681149] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 796.681149] env[61987]: value = "task-1061644" [ 796.681149] env[61987]: _type = "Task" [ 796.681149] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.689945] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061644, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.756647] env[61987]: DEBUG nova.network.neutron [-] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.932245] env[61987]: DEBUG oslo_concurrency.lockutils [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] Releasing lock "refresh_cache-10354ba0-5a64-4174-9c61-152d54a5dfcc" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.932245] env[61987]: DEBUG nova.compute.manager [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Received event network-vif-plugged-a3d8a0f6-c3dc-42d9-b4e1-903a299b601e {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 796.932245] env[61987]: DEBUG oslo_concurrency.lockutils [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] Acquiring lock "4fd8fc17-fb13-42e7-9879-9f120fdcb330-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.932245] env[61987]: DEBUG oslo_concurrency.lockutils [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] Lock "4fd8fc17-fb13-42e7-9879-9f120fdcb330-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.932245] env[61987]: DEBUG oslo_concurrency.lockutils [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] Lock "4fd8fc17-fb13-42e7-9879-9f120fdcb330-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.932245] env[61987]: DEBUG nova.compute.manager [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] No waiting events found dispatching network-vif-plugged-a3d8a0f6-c3dc-42d9-b4e1-903a299b601e {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 796.932245] env[61987]: WARNING nova.compute.manager [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Received unexpected event network-vif-plugged-a3d8a0f6-c3dc-42d9-b4e1-903a299b601e for instance with vm_state building and task_state spawning. [ 796.932245] env[61987]: DEBUG nova.compute.manager [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Received event network-changed-a3d8a0f6-c3dc-42d9-b4e1-903a299b601e {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 796.932585] env[61987]: DEBUG nova.compute.manager [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Refreshing instance network info cache due to event network-changed-a3d8a0f6-c3dc-42d9-b4e1-903a299b601e. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 796.932585] env[61987]: DEBUG oslo_concurrency.lockutils [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] Acquiring lock "refresh_cache-4fd8fc17-fb13-42e7-9879-9f120fdcb330" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.932818] env[61987]: DEBUG oslo_concurrency.lockutils [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] Acquired lock "refresh_cache-4fd8fc17-fb13-42e7-9879-9f120fdcb330" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.933614] env[61987]: DEBUG nova.network.neutron [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Refreshing network info cache for port a3d8a0f6-c3dc-42d9-b4e1-903a299b601e {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 796.934300] env[61987]: DEBUG oslo_concurrency.lockutils [req-0e41a055-ddec-4fba-b221-0ceb8542ffdf req-19ae98b8-e3bd-4dcf-b3ee-2a62c488a482 service nova] Acquired lock "refresh_cache-10354ba0-5a64-4174-9c61-152d54a5dfcc" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.934416] env[61987]: DEBUG nova.network.neutron [req-0e41a055-ddec-4fba-b221-0ceb8542ffdf req-19ae98b8-e3bd-4dcf-b3ee-2a62c488a482 service nova] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Refreshing network info cache for port b9f8f891-82e4-440d-bf4d-5b3c70c910e6 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 796.972046] env[61987]: DEBUG nova.network.neutron [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Successfully created port: e7d811c0-d77e-448d-86fe-acdb61c04284 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 797.012565] env[61987]: DEBUG oslo_vmware.api [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061643, 'name': PowerOffVM_Task, 'duration_secs': 0.317813} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.012565] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 797.012565] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 797.012565] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2ff403c2-f7bb-4cbc-b3c6-f76048c2ccd4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.096215] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 797.097306] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 797.097744] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Deleting the datastore file [datastore2] 10354ba0-5a64-4174-9c61-152d54a5dfcc {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 797.100641] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 797.101733] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8968cadf-f46e-41b8-b20e-a2245e7c3d46 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.105041] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3d6fb27d-b871-4028-93e3-68283a496a2e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.115259] env[61987]: DEBUG oslo_vmware.api [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for the task: (returnval){ [ 797.115259] env[61987]: value = "task-1061647" [ 797.115259] env[61987]: _type = "Task" [ 797.115259] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.115259] env[61987]: DEBUG oslo_vmware.api [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 797.115259] env[61987]: value = "task-1061646" [ 797.115259] env[61987]: _type = "Task" [ 797.115259] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.126253] env[61987]: DEBUG oslo_vmware.api [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061646, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.129747] env[61987]: DEBUG oslo_vmware.api [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061647, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.148279] env[61987]: DEBUG nova.objects.base [None req-6a08f2fb-6443-43ca-aa8c-0a34883d3e6c tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61987) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 797.185021] env[61987]: DEBUG nova.scheduler.client.report [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 797.200366] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061644, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070713} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.200366] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 797.201306] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db8abbb-5a70-4871-9b1f-1037ed567780 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.234107] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] 4fd8fc17-fb13-42e7-9879-9f120fdcb330/4fd8fc17-fb13-42e7-9879-9f120fdcb330.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 797.234996] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e46a07bf-06b2-43d8-9506-b2d4461384ff {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.259405] env[61987]: INFO nova.compute.manager [-] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Took 2.10 seconds to deallocate network for instance. [ 797.264630] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 797.264630] env[61987]: value = "task-1061648" [ 797.264630] env[61987]: _type = "Task" [ 797.264630] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.277543] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061648, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.606791] env[61987]: DEBUG nova.compute.manager [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 797.632715] env[61987]: DEBUG oslo_vmware.api [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061647, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180106} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.638107] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 797.638218] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 797.638470] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 797.638702] env[61987]: INFO nova.compute.manager [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Took 1.17 seconds to destroy the instance on the hypervisor. [ 797.639012] env[61987]: DEBUG oslo.service.loopingcall [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.639329] env[61987]: DEBUG oslo_vmware.api [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061646, 'name': PowerOffVM_Task, 'duration_secs': 0.226093} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.641698] env[61987]: DEBUG nova.compute.manager [-] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 797.641889] env[61987]: DEBUG nova.network.neutron [-] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 797.643512] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 797.643761] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Updating instance '8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2' progress to 17 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 797.664052] env[61987]: DEBUG nova.virt.hardware [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 797.664052] env[61987]: DEBUG nova.virt.hardware [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 797.664052] env[61987]: DEBUG nova.virt.hardware [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 797.664052] env[61987]: DEBUG nova.virt.hardware [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 797.664052] env[61987]: DEBUG nova.virt.hardware [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 797.664052] env[61987]: DEBUG nova.virt.hardware [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 797.664052] env[61987]: DEBUG nova.virt.hardware [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 797.664052] env[61987]: DEBUG nova.virt.hardware [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 797.664052] env[61987]: DEBUG nova.virt.hardware [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 797.664052] env[61987]: DEBUG nova.virt.hardware [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 797.664052] env[61987]: DEBUG nova.virt.hardware [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 797.664052] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a751f34-54e7-4d61-9bb5-f80e4a5da812 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.673069] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3641edd-6d35-4b74-a0fe-5b94cffa196c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.694552] env[61987]: DEBUG oslo_concurrency.lockutils [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.131s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.697118] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.890s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.697649] env[61987]: DEBUG nova.objects.instance [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Lazy-loading 'resources' on Instance uuid bb7243dc-a950-4d3e-9b17-944e57cdd56f {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 797.723528] env[61987]: INFO nova.scheduler.client.report [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Deleted allocations for instance 1afcffda-0bd2-4cbe-8cfb-12a91bb50975 [ 797.775904] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061648, 'name': ReconfigVM_Task, 'duration_secs': 0.305536} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.776312] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Reconfigured VM instance instance-0000003e to attach disk [datastore1] 4fd8fc17-fb13-42e7-9879-9f120fdcb330/4fd8fc17-fb13-42e7-9879-9f120fdcb330.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 797.777303] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-27908fd2-689a-4d56-acc3-513d729fb8b6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.783597] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 797.783597] env[61987]: value = "task-1061649" [ 797.783597] env[61987]: _type = "Task" [ 797.783597] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.798123] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061649, 'name': Rename_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.838217] env[61987]: INFO nova.compute.manager [None req-acb92527-6504-4fe5-bca2-a703c7bd5326 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Took 0.58 seconds to detach 1 volumes for instance. [ 797.839394] env[61987]: DEBUG nova.compute.manager [None req-acb92527-6504-4fe5-bca2-a703c7bd5326 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Deleting volume: a8d52043-effe-4e4a-9a52-3ed5b5050a94 {{(pid=61987) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3281}} [ 797.978963] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7dbe2253-a13a-4b96-8b92-a9f8e382928c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "190a1100-a3c1-4292-be2a-015627f0efa4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.979278] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7dbe2253-a13a-4b96-8b92-a9f8e382928c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "190a1100-a3c1-4292-be2a-015627f0efa4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.152580] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 798.152853] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 798.153032] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 798.153552] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 798.153766] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 798.154019] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 798.154317] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 798.154527] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 798.154741] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 798.155095] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 798.155307] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 798.167145] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-042324d9-3128-48c5-8d3a-0a5fb7851f77 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.179849] env[61987]: DEBUG nova.network.neutron [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Updated VIF entry in instance network info cache for port a3d8a0f6-c3dc-42d9-b4e1-903a299b601e. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 798.180154] env[61987]: DEBUG nova.network.neutron [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Updating instance_info_cache with network_info: [{"id": "a3d8a0f6-c3dc-42d9-b4e1-903a299b601e", "address": "fa:16:3e:86:34:de", "network": {"id": "8dfc473a-6d57-4169-9338-49a9565bb93d", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-422412073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c3d31c138df453486a2b1a6b70a84d4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa3d8a0f6-c3", "ovs_interfaceid": "a3d8a0f6-c3dc-42d9-b4e1-903a299b601e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.187739] env[61987]: DEBUG oslo_vmware.api [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 798.187739] env[61987]: value = "task-1061651" [ 798.187739] env[61987]: _type = "Task" [ 798.187739] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.197452] env[61987]: DEBUG oslo_vmware.api [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061651, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.210428] env[61987]: DEBUG nova.network.neutron [req-0e41a055-ddec-4fba-b221-0ceb8542ffdf req-19ae98b8-e3bd-4dcf-b3ee-2a62c488a482 service nova] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Updated VIF entry in instance network info cache for port b9f8f891-82e4-440d-bf4d-5b3c70c910e6. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 798.214014] env[61987]: DEBUG nova.network.neutron [req-0e41a055-ddec-4fba-b221-0ceb8542ffdf req-19ae98b8-e3bd-4dcf-b3ee-2a62c488a482 service nova] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Updating instance_info_cache with network_info: [{"id": "b9f8f891-82e4-440d-bf4d-5b3c70c910e6", "address": "fa:16:3e:0e:8f:0f", "network": {"id": "90535a3b-74ac-4bdc-a5ea-78272b01cab6", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1880668508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d8a522ba9bf422db98ca69c8d92b373", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb9f8f891-82", "ovs_interfaceid": "b9f8f891-82e4-440d-bf4d-5b3c70c910e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.238301] env[61987]: DEBUG oslo_concurrency.lockutils [None req-af9b11dc-fa60-4905-963b-58e24fd7d341 tempest-ServersAdmin275Test-864401277 tempest-ServersAdmin275Test-864401277-project-member] Lock "1afcffda-0bd2-4cbe-8cfb-12a91bb50975" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.886s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.300839] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061649, 'name': Rename_Task, 'duration_secs': 0.159675} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.300839] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 798.300839] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-30ab014f-043d-49a1-b96e-98d3cb252e7a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.310697] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 798.310697] env[61987]: value = "task-1061652" [ 798.310697] env[61987]: _type = "Task" [ 798.310697] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.321437] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061652, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.390751] env[61987]: DEBUG oslo_concurrency.lockutils [None req-acb92527-6504-4fe5-bca2-a703c7bd5326 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.454218] env[61987]: DEBUG nova.network.neutron [None req-6a08f2fb-6443-43ca-aa8c-0a34883d3e6c tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Updating instance_info_cache with network_info: [{"id": "6aa03b99-d58f-461e-b2a8-c4552c1d418a", "address": "fa:16:3e:5e:3e:15", "network": {"id": "a9093c43-6bc4-45b1-bdcf-a434759300bd", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-355385599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.128", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7efd6c745f547d3803d5a6184547bda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac563aa7-6d7c-4bd5-9241-7b3e11b8f22d", "external-id": "nsx-vlan-transportzone-730", "segmentation_id": 730, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6aa03b99-d5", "ovs_interfaceid": "6aa03b99-d58f-461e-b2a8-c4552c1d418a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.482614] env[61987]: DEBUG nova.compute.manager [None req-7dbe2253-a13a-4b96-8b92-a9f8e382928c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 190a1100-a3c1-4292-be2a-015627f0efa4] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 798.683596] env[61987]: DEBUG oslo_concurrency.lockutils [req-6b55bcd8-874b-4680-8baf-09e02de6bb45 req-303a05d6-9c46-4a38-adb6-0ec92fecba1c service nova] Releasing lock "refresh_cache-4fd8fc17-fb13-42e7-9879-9f120fdcb330" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.701079] env[61987]: DEBUG oslo_vmware.api [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061651, 'name': ReconfigVM_Task, 'duration_secs': 0.233753} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.701473] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Updating instance '8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2' progress to 33 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 798.716857] env[61987]: DEBUG oslo_concurrency.lockutils [req-0e41a055-ddec-4fba-b221-0ceb8542ffdf req-19ae98b8-e3bd-4dcf-b3ee-2a62c488a482 service nova] Releasing lock "refresh_cache-10354ba0-5a64-4174-9c61-152d54a5dfcc" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.732113] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3310cf2e-b7d9-47e9-b06a-5530ca3b3756 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.740595] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-362735de-e233-44bf-a9c5-e48a040309bb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.788514] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee1fd2e7-72b4-4d3c-929b-3fde9a71cb25 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.799574] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6239471b-bddd-4bed-80ac-32036da7d3a1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.812018] env[61987]: DEBUG nova.compute.manager [req-f82c3520-18e8-4d09-b598-14aeb8eb2f80 req-fa2e5b2f-fe8d-4f9a-be07-2326ff1e8d97 service nova] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Received event network-vif-deleted-021ce2e6-14d2-4f08-b654-6770827da9ee {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 798.812018] env[61987]: DEBUG nova.compute.manager [req-f82c3520-18e8-4d09-b598-14aeb8eb2f80 req-fa2e5b2f-fe8d-4f9a-be07-2326ff1e8d97 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Received event network-changed-39309545-c724-44c7-932e-7784c9e31fa7 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 798.812018] env[61987]: DEBUG nova.compute.manager [req-f82c3520-18e8-4d09-b598-14aeb8eb2f80 req-fa2e5b2f-fe8d-4f9a-be07-2326ff1e8d97 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Refreshing instance network info cache due to event network-changed-39309545-c724-44c7-932e-7784c9e31fa7. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 798.812018] env[61987]: DEBUG oslo_concurrency.lockutils [req-f82c3520-18e8-4d09-b598-14aeb8eb2f80 req-fa2e5b2f-fe8d-4f9a-be07-2326ff1e8d97 service nova] Acquiring lock "refresh_cache-4edd0159-6db1-41bd-a616-31a34e984059" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.812018] env[61987]: DEBUG oslo_concurrency.lockutils [req-f82c3520-18e8-4d09-b598-14aeb8eb2f80 req-fa2e5b2f-fe8d-4f9a-be07-2326ff1e8d97 service nova] Acquired lock "refresh_cache-4edd0159-6db1-41bd-a616-31a34e984059" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.812018] env[61987]: DEBUG nova.network.neutron [req-f82c3520-18e8-4d09-b598-14aeb8eb2f80 req-fa2e5b2f-fe8d-4f9a-be07-2326ff1e8d97 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Refreshing network info cache for port 39309545-c724-44c7-932e-7784c9e31fa7 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 798.822929] env[61987]: DEBUG nova.compute.provider_tree [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 798.835058] env[61987]: DEBUG oslo_vmware.api [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061652, 'name': PowerOnVM_Task, 'duration_secs': 0.4935} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.835196] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 798.835393] env[61987]: INFO nova.compute.manager [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Took 7.43 seconds to spawn the instance on the hypervisor. [ 798.835594] env[61987]: DEBUG nova.compute.manager [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 798.836495] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd59c94-0394-42e5-a96c-f9400263d3aa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.923405] env[61987]: DEBUG nova.network.neutron [-] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.956689] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a08f2fb-6443-43ca-aa8c-0a34883d3e6c tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Releasing lock "refresh_cache-ad40937f-d5a9-4708-8b81-06087d38d765" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.989037] env[61987]: DEBUG nova.compute.manager [None req-7dbe2253-a13a-4b96-8b92-a9f8e382928c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 190a1100-a3c1-4292-be2a-015627f0efa4] Instance disappeared before build. {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2445}} [ 799.209973] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 799.210565] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 799.210862] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 799.211245] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 799.211536] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 799.211833] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 799.212192] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 799.212496] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 799.212783] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 799.213091] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 799.213412] env[61987]: DEBUG nova.virt.hardware [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 799.225317] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Reconfiguring VM instance instance-00000034 to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 799.225778] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-887ae7ca-d8cb-41ac-a2b4-01e69f0b56ef {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.260082] env[61987]: DEBUG oslo_vmware.api [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 799.260082] env[61987]: value = "task-1061653" [ 799.260082] env[61987]: _type = "Task" [ 799.260082] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.274665] env[61987]: DEBUG oslo_vmware.api [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061653, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.330351] env[61987]: DEBUG nova.scheduler.client.report [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 799.357122] env[61987]: INFO nova.compute.manager [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Took 40.50 seconds to build instance. [ 799.427164] env[61987]: INFO nova.compute.manager [-] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Took 1.78 seconds to deallocate network for instance. [ 799.505440] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7dbe2253-a13a-4b96-8b92-a9f8e382928c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "190a1100-a3c1-4292-be2a-015627f0efa4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 1.526s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.689054] env[61987]: DEBUG nova.network.neutron [req-f82c3520-18e8-4d09-b598-14aeb8eb2f80 req-fa2e5b2f-fe8d-4f9a-be07-2326ff1e8d97 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Updated VIF entry in instance network info cache for port 39309545-c724-44c7-932e-7784c9e31fa7. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 799.689411] env[61987]: DEBUG nova.network.neutron [req-f82c3520-18e8-4d09-b598-14aeb8eb2f80 req-fa2e5b2f-fe8d-4f9a-be07-2326ff1e8d97 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Updating instance_info_cache with network_info: [{"id": "39309545-c724-44c7-932e-7784c9e31fa7", "address": "fa:16:3e:fd:80:60", "network": {"id": "90535a3b-74ac-4bdc-a5ea-78272b01cab6", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1880668508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d8a522ba9bf422db98ca69c8d92b373", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39309545-c7", "ovs_interfaceid": "39309545-c724-44c7-932e-7784c9e31fa7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.771998] env[61987]: DEBUG nova.network.neutron [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Successfully updated port: e7d811c0-d77e-448d-86fe-acdb61c04284 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 799.781298] env[61987]: DEBUG oslo_vmware.api [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061653, 'name': ReconfigVM_Task, 'duration_secs': 0.194162} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.781898] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Reconfigured VM instance instance-00000034 to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 799.783222] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb86c979-c8f4-449f-a4e6-0bfa62678f07 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.812489] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2/8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 799.813553] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5cf82dc6-e743-4222-b84e-f02cc6742cd7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.834864] env[61987]: DEBUG oslo_vmware.api [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 799.834864] env[61987]: value = "task-1061654" [ 799.834864] env[61987]: _type = "Task" [ 799.834864] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.836076] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.139s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.842412] env[61987]: DEBUG oslo_concurrency.lockutils [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 30.130s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.849359] env[61987]: DEBUG oslo_vmware.api [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061654, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.860017] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f590bc10-3a20-4fe2-9a7d-fa830c890aa8 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "4fd8fc17-fb13-42e7-9879-9f120fdcb330" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.644s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.864356] env[61987]: INFO nova.scheduler.client.report [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Deleted allocations for instance bb7243dc-a950-4d3e-9b17-944e57cdd56f [ 799.934613] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.967784] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a08f2fb-6443-43ca-aa8c-0a34883d3e6c tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 799.967784] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b4b7e25a-c6f5-4e72-a12a-f64892d67dcf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.974484] env[61987]: DEBUG oslo_vmware.api [None req-6a08f2fb-6443-43ca-aa8c-0a34883d3e6c tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 799.974484] env[61987]: value = "task-1061655" [ 799.974484] env[61987]: _type = "Task" [ 799.974484] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.984380] env[61987]: DEBUG oslo_vmware.api [None req-6a08f2fb-6443-43ca-aa8c-0a34883d3e6c tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061655, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.192852] env[61987]: DEBUG oslo_concurrency.lockutils [req-f82c3520-18e8-4d09-b598-14aeb8eb2f80 req-fa2e5b2f-fe8d-4f9a-be07-2326ff1e8d97 service nova] Releasing lock "refresh_cache-4edd0159-6db1-41bd-a616-31a34e984059" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.284625] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "refresh_cache-1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.284625] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "refresh_cache-1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.284625] env[61987]: DEBUG nova.network.neutron [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 800.317476] env[61987]: DEBUG nova.compute.manager [req-c0cceef5-2401-4e04-a512-139cb9cfefae req-ca7328e6-136a-4fca-aed2-9fb1b7418292 service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Received event network-vif-plugged-e7d811c0-d77e-448d-86fe-acdb61c04284 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 800.318231] env[61987]: DEBUG oslo_concurrency.lockutils [req-c0cceef5-2401-4e04-a512-139cb9cfefae req-ca7328e6-136a-4fca-aed2-9fb1b7418292 service nova] Acquiring lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.318364] env[61987]: DEBUG oslo_concurrency.lockutils [req-c0cceef5-2401-4e04-a512-139cb9cfefae req-ca7328e6-136a-4fca-aed2-9fb1b7418292 service nova] Lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.318577] env[61987]: DEBUG oslo_concurrency.lockutils [req-c0cceef5-2401-4e04-a512-139cb9cfefae req-ca7328e6-136a-4fca-aed2-9fb1b7418292 service nova] Lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.318710] env[61987]: DEBUG nova.compute.manager [req-c0cceef5-2401-4e04-a512-139cb9cfefae req-ca7328e6-136a-4fca-aed2-9fb1b7418292 service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] No waiting events found dispatching network-vif-plugged-e7d811c0-d77e-448d-86fe-acdb61c04284 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 800.318883] env[61987]: WARNING nova.compute.manager [req-c0cceef5-2401-4e04-a512-139cb9cfefae req-ca7328e6-136a-4fca-aed2-9fb1b7418292 service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Received unexpected event network-vif-plugged-e7d811c0-d77e-448d-86fe-acdb61c04284 for instance with vm_state building and task_state spawning. [ 800.344716] env[61987]: DEBUG oslo_vmware.api [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061654, 'name': ReconfigVM_Task, 'duration_secs': 0.33332} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.345049] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Reconfigured VM instance instance-00000034 to attach disk [datastore2] 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2/8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 800.345348] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Updating instance '8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2' progress to 50 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 800.349301] env[61987]: DEBUG nova.objects.instance [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lazy-loading 'migration_context' on Instance uuid 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 800.374034] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5b76baab-9f6c-4f82-b3a2-b249692b01a8 tempest-VolumesAdminNegativeTest-2029357998 tempest-VolumesAdminNegativeTest-2029357998-project-member] Lock "bb7243dc-a950-4d3e-9b17-944e57cdd56f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.280s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.378174] env[61987]: DEBUG nova.compute.manager [req-e6dbded7-3a9d-46c6-92dd-40505a2ee1f5 req-4f089f52-9556-428f-b9dc-e24cbb445b17 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Received event network-changed-39309545-c724-44c7-932e-7784c9e31fa7 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 800.379030] env[61987]: DEBUG nova.compute.manager [req-e6dbded7-3a9d-46c6-92dd-40505a2ee1f5 req-4f089f52-9556-428f-b9dc-e24cbb445b17 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Refreshing instance network info cache due to event network-changed-39309545-c724-44c7-932e-7784c9e31fa7. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 800.379030] env[61987]: DEBUG oslo_concurrency.lockutils [req-e6dbded7-3a9d-46c6-92dd-40505a2ee1f5 req-4f089f52-9556-428f-b9dc-e24cbb445b17 service nova] Acquiring lock "refresh_cache-4edd0159-6db1-41bd-a616-31a34e984059" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.379030] env[61987]: DEBUG oslo_concurrency.lockutils [req-e6dbded7-3a9d-46c6-92dd-40505a2ee1f5 req-4f089f52-9556-428f-b9dc-e24cbb445b17 service nova] Acquired lock "refresh_cache-4edd0159-6db1-41bd-a616-31a34e984059" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.379030] env[61987]: DEBUG nova.network.neutron [req-e6dbded7-3a9d-46c6-92dd-40505a2ee1f5 req-4f089f52-9556-428f-b9dc-e24cbb445b17 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Refreshing network info cache for port 39309545-c724-44c7-932e-7784c9e31fa7 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 800.489623] env[61987]: DEBUG oslo_vmware.api [None req-6a08f2fb-6443-43ca-aa8c-0a34883d3e6c tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061655, 'name': PowerOnVM_Task, 'duration_secs': 0.40483} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.489623] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a08f2fb-6443-43ca-aa8c-0a34883d3e6c tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 800.489623] env[61987]: DEBUG nova.compute.manager [None req-6a08f2fb-6443-43ca-aa8c-0a34883d3e6c tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 800.489623] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df6f24b-2d7a-44b3-9065-ce19f0fa70a5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.807028] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "1961c2c7-4400-4559-9f7c-bec8e8e263b0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.807028] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "1961c2c7-4400-4559-9f7c-bec8e8e263b0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.840014] env[61987]: DEBUG nova.network.neutron [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 800.863011] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d3e7456-679a-4b5d-9707-eaa1158629a6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.896053] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ef867db-7999-48d5-8033-465748f9edeb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.922105] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Updating instance '8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2' progress to 67 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 801.083109] env[61987]: DEBUG nova.network.neutron [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Updating instance_info_cache with network_info: [{"id": "e7d811c0-d77e-448d-86fe-acdb61c04284", "address": "fa:16:3e:b0:6d:8c", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7d811c0-d7", "ovs_interfaceid": "e7d811c0-d77e-448d-86fe-acdb61c04284", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.202899] env[61987]: DEBUG nova.compute.manager [req-01cfc752-4362-4669-99dd-d4b157f2d15e req-bb8f7701-b580-4b7e-99fe-5e93c5d77841 service nova] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Received event network-vif-deleted-b9f8f891-82e4-440d-bf4d-5b3c70c910e6 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 801.246503] env[61987]: DEBUG nova.network.neutron [req-e6dbded7-3a9d-46c6-92dd-40505a2ee1f5 req-4f089f52-9556-428f-b9dc-e24cbb445b17 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Updated VIF entry in instance network info cache for port 39309545-c724-44c7-932e-7784c9e31fa7. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 801.246855] env[61987]: DEBUG nova.network.neutron [req-e6dbded7-3a9d-46c6-92dd-40505a2ee1f5 req-4f089f52-9556-428f-b9dc-e24cbb445b17 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Updating instance_info_cache with network_info: [{"id": "39309545-c724-44c7-932e-7784c9e31fa7", "address": "fa:16:3e:fd:80:60", "network": {"id": "90535a3b-74ac-4bdc-a5ea-78272b01cab6", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-1880668508-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8d8a522ba9bf422db98ca69c8d92b373", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60567ee6-01d0-4b16-9c7a-4a896827d6eb", "external-id": "nsx-vlan-transportzone-28", "segmentation_id": 28, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap39309545-c7", "ovs_interfaceid": "39309545-c724-44c7-932e-7784c9e31fa7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.271572] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf1f76a-1629-4209-af89-0a8209096a63 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.278652] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ec62fa3-5350-4b50-97dc-e34715d9544d tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquiring lock "10354ba0-5a64-4174-9c61-152d54a5dfcc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.280010] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6530706-f30e-469d-8c19-0239d1eaa403 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.312823] env[61987]: DEBUG nova.compute.manager [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 801.316423] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea85a77-7cf6-436e-b4bc-880e01f00c76 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.332520] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c952171-c2bc-40a4-b427-ecbd3181f4cb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.348138] env[61987]: DEBUG nova.compute.provider_tree [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.472573] env[61987]: DEBUG nova.network.neutron [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Port 55e46e63-9076-400e-a73a-19fc4bb21bec binding to destination host cpu-1 is already ACTIVE {{(pid=61987) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 801.585680] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "refresh_cache-1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.586043] env[61987]: DEBUG nova.compute.manager [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Instance network_info: |[{"id": "e7d811c0-d77e-448d-86fe-acdb61c04284", "address": "fa:16:3e:b0:6d:8c", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7d811c0-d7", "ovs_interfaceid": "e7d811c0-d77e-448d-86fe-acdb61c04284", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 801.586586] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:6d:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1c8fdf9-970c-4ae0-b6d9-f1015196b552', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e7d811c0-d77e-448d-86fe-acdb61c04284', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 801.594029] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Creating folder: Project (d2afdc9bddba4ccab2b8fdb1f31da2b6). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 801.594418] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-66370d9b-d97b-4a93-ae82-95a339452f23 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.606410] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Created folder: Project (d2afdc9bddba4ccab2b8fdb1f31da2b6) in parent group-v234219. [ 801.606632] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Creating folder: Instances. Parent ref: group-v234331. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 801.606929] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-627b8e50-488e-4dc3-91ee-90cf1d90922b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.617988] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Created folder: Instances in parent group-v234331. [ 801.618334] env[61987]: DEBUG oslo.service.loopingcall [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 801.618671] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 801.618942] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-34f9fc45-d748-4282-bc3b-4fd1d440c96a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.639748] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 801.639748] env[61987]: value = "task-1061658" [ 801.639748] env[61987]: _type = "Task" [ 801.639748] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.647578] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061658, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.749617] env[61987]: DEBUG oslo_concurrency.lockutils [req-e6dbded7-3a9d-46c6-92dd-40505a2ee1f5 req-4f089f52-9556-428f-b9dc-e24cbb445b17 service nova] Releasing lock "refresh_cache-4edd0159-6db1-41bd-a616-31a34e984059" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.844175] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.851642] env[61987]: DEBUG nova.scheduler.client.report [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 802.049343] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "f937a5ec-a1d7-41d8-b998-fa18f545b304" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.049589] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "f937a5ec-a1d7-41d8-b998-fa18f545b304" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.155065] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061658, 'name': CreateVM_Task, 'duration_secs': 0.479949} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.155580] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 802.156066] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.156236] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.156590] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 802.156860] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f982b12-1dfa-4160-8481-34b751fc907b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.162032] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 802.162032] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52a7b5ad-60ec-b706-a62d-ad15ec6b2ebf" [ 802.162032] env[61987]: _type = "Task" [ 802.162032] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.169552] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52a7b5ad-60ec-b706-a62d-ad15ec6b2ebf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.502219] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.503160] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.503559] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.553567] env[61987]: DEBUG nova.compute.utils [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 802.677026] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52a7b5ad-60ec-b706-a62d-ad15ec6b2ebf, 'name': SearchDatastore_Task, 'duration_secs': 0.020501} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.677026] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.677026] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 802.677702] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.678162] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.678640] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 802.679192] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-356192f8-3995-4229-ab44-f350653f33dd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.692246] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 802.692246] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 802.693281] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-538fcb7b-29a6-4d70-8c9c-d15611e5f625 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.702313] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 802.702313] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]522e8219-58e5-aa2b-1282-71e49435e487" [ 802.702313] env[61987]: _type = "Task" [ 802.702313] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.708477] env[61987]: DEBUG nova.compute.manager [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 802.710313] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c9febf-cc9e-4ea6-b3c3-b857f7e4a18d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.722321] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]522e8219-58e5-aa2b-1282-71e49435e487, 'name': SearchDatastore_Task, 'duration_secs': 0.010998} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.726976] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab891357-10a4-497a-a154-9c0c8eeb131e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.733184] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 802.733184] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]529276fd-cc02-1bb4-8530-2c440dbaf4c2" [ 802.733184] env[61987]: _type = "Task" [ 802.733184] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.741589] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]529276fd-cc02-1bb4-8530-2c440dbaf4c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.847470] env[61987]: DEBUG nova.compute.manager [req-a12f2948-915c-43ef-a18e-160accff5988 req-dd84f22f-fabf-4f2c-bfbb-9c10d659734f service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Received event network-changed-e7d811c0-d77e-448d-86fe-acdb61c04284 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 802.848017] env[61987]: DEBUG nova.compute.manager [req-a12f2948-915c-43ef-a18e-160accff5988 req-dd84f22f-fabf-4f2c-bfbb-9c10d659734f service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Refreshing instance network info cache due to event network-changed-e7d811c0-d77e-448d-86fe-acdb61c04284. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 802.848257] env[61987]: DEBUG oslo_concurrency.lockutils [req-a12f2948-915c-43ef-a18e-160accff5988 req-dd84f22f-fabf-4f2c-bfbb-9c10d659734f service nova] Acquiring lock "refresh_cache-1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.848440] env[61987]: DEBUG oslo_concurrency.lockutils [req-a12f2948-915c-43ef-a18e-160accff5988 req-dd84f22f-fabf-4f2c-bfbb-9c10d659734f service nova] Acquired lock "refresh_cache-1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.848622] env[61987]: DEBUG nova.network.neutron [req-a12f2948-915c-43ef-a18e-160accff5988 req-dd84f22f-fabf-4f2c-bfbb-9c10d659734f service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Refreshing network info cache for port e7d811c0-d77e-448d-86fe-acdb61c04284 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 802.867743] env[61987]: DEBUG oslo_concurrency.lockutils [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.025s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.887296] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.254s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.887840] env[61987]: DEBUG nova.objects.instance [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Lazy-loading 'resources' on Instance uuid 466a5f5a-f48a-4921-8bcb-44a1840c8141 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 803.056800] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "f937a5ec-a1d7-41d8-b998-fa18f545b304" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.231180] env[61987]: INFO nova.compute.manager [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] instance snapshotting [ 803.234331] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd44aff7-3426-4f4e-a3ff-44ab2872891a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.247045] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]529276fd-cc02-1bb4-8530-2c440dbaf4c2, 'name': SearchDatastore_Task, 'duration_secs': 0.013165} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.259636] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.259983] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 1f29d5c1-bd01-47dc-9dcf-4e89791120d0/1f29d5c1-bd01-47dc-9dcf-4e89791120d0.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 803.260457] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3003923b-21b9-4cdd-acc9-28d977c4baf6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.263115] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5930da6-1c01-46ea-acbc-ceaec5004bd1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.273850] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 803.273850] env[61987]: value = "task-1061659" [ 803.273850] env[61987]: _type = "Task" [ 803.273850] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.282148] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061659, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.565676] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "refresh_cache-8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.565879] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired lock "refresh_cache-8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.566118] env[61987]: DEBUG nova.network.neutron [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 803.776624] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Creating Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 803.777353] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9082e746-b736-4242-b860-643b79a64213 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.784799] env[61987]: DEBUG nova.network.neutron [req-a12f2948-915c-43ef-a18e-160accff5988 req-dd84f22f-fabf-4f2c-bfbb-9c10d659734f service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Updated VIF entry in instance network info cache for port e7d811c0-d77e-448d-86fe-acdb61c04284. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 803.785210] env[61987]: DEBUG nova.network.neutron [req-a12f2948-915c-43ef-a18e-160accff5988 req-dd84f22f-fabf-4f2c-bfbb-9c10d659734f service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Updating instance_info_cache with network_info: [{"id": "e7d811c0-d77e-448d-86fe-acdb61c04284", "address": "fa:16:3e:b0:6d:8c", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7d811c0-d7", "ovs_interfaceid": "e7d811c0-d77e-448d-86fe-acdb61c04284", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.800068] env[61987]: DEBUG oslo_vmware.api [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 803.800068] env[61987]: value = "task-1061660" [ 803.800068] env[61987]: _type = "Task" [ 803.800068] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.800407] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061659, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.813312] env[61987]: DEBUG oslo_vmware.api [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061660, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.879247] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a5bc9fe-a0df-4a0a-9e5b-ac3d00c8c710 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.887124] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdc04523-bb01-40f6-831a-de8b47fa645f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.922467] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ae0be17-fb3c-48fb-971e-893f063d2da2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.936089] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-001297a4-5546-40cd-ad9c-431502785a00 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.957959] env[61987]: DEBUG nova.compute.provider_tree [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.015426] env[61987]: DEBUG oslo_vmware.rw_handles [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293ff7f-4a7a-78b8-22eb-53d6116ccfe1/disk-0.vmdk. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 804.016577] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83f5cb8-bdb4-4ff0-aa55-cb45a93491d5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.023288] env[61987]: DEBUG oslo_vmware.rw_handles [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293ff7f-4a7a-78b8-22eb-53d6116ccfe1/disk-0.vmdk is in state: ready. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 804.023423] env[61987]: ERROR oslo_vmware.rw_handles [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293ff7f-4a7a-78b8-22eb-53d6116ccfe1/disk-0.vmdk due to incomplete transfer. [ 804.023650] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5852f56b-c098-44ef-ba84-32f7baf0687a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.030363] env[61987]: DEBUG oslo_vmware.rw_handles [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5293ff7f-4a7a-78b8-22eb-53d6116ccfe1/disk-0.vmdk. {{(pid=61987) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 804.030587] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Uploaded image 92117f13-9c42-454c-87bb-8c7e90801e79 to the Glance image server {{(pid=61987) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 804.033408] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Destroying the VM {{(pid=61987) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 804.033681] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-caa54ab9-b2ec-4f5d-ba30-c83ab68a097d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.040399] env[61987]: DEBUG oslo_vmware.api [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 804.040399] env[61987]: value = "task-1061661" [ 804.040399] env[61987]: _type = "Task" [ 804.040399] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.048475] env[61987]: DEBUG oslo_vmware.api [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061661, 'name': Destroy_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.159695] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "f937a5ec-a1d7-41d8-b998-fa18f545b304" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.159978] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "f937a5ec-a1d7-41d8-b998-fa18f545b304" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.160241] env[61987]: INFO nova.compute.manager [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Attaching volume 41cab2ad-6048-4d47-a5db-77ff38f4f0f0 to /dev/sdb [ 804.209456] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4981d9bb-7719-47c8-9286-71da60e0ce52 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.217214] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46c6c28a-9171-41c6-9f9e-b6f1231fe545 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.232031] env[61987]: DEBUG nova.virt.block_device [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Updating existing volume attachment record: ffca4aa5-68ec-40da-bc56-b5bdbaa384e9 {{(pid=61987) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 804.287302] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061659, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.618448} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.287700] env[61987]: DEBUG oslo_concurrency.lockutils [req-a12f2948-915c-43ef-a18e-160accff5988 req-dd84f22f-fabf-4f2c-bfbb-9c10d659734f service nova] Releasing lock "refresh_cache-1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.288090] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 1f29d5c1-bd01-47dc-9dcf-4e89791120d0/1f29d5c1-bd01-47dc-9dcf-4e89791120d0.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 804.288321] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 804.288578] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8422b312-b43b-4f52-814e-fdbdfe62bae4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.294540] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 804.294540] env[61987]: value = "task-1061662" [ 804.294540] env[61987]: _type = "Task" [ 804.294540] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.302718] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061662, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.311125] env[61987]: DEBUG oslo_vmware.api [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061660, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.459242] env[61987]: INFO nova.compute.manager [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Swapping old allocation on dict_keys(['8db744e5-1e14-4e98-9818-695ea4b7dfd0']) held by migration 0f243de7-62e3-4127-8bbb-f1c7b76ff14d for instance [ 804.463585] env[61987]: DEBUG nova.scheduler.client.report [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 804.502100] env[61987]: DEBUG nova.scheduler.client.report [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Overwriting current allocation {'allocations': {'8db744e5-1e14-4e98-9818-695ea4b7dfd0': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 84}}, 'project_id': 'd4e086581d9e41b197f84dac49aab72e', 'user_id': 'e6b9b9da9b9248d3bffdb97582f0b1fe', 'consumer_generation': 1} on consumer 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6 {{(pid=61987) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 804.523631] env[61987]: DEBUG nova.network.neutron [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Updating instance_info_cache with network_info: [{"id": "55e46e63-9076-400e-a73a-19fc4bb21bec", "address": "fa:16:3e:09:61:0c", "network": {"id": "59d8d7d6-fd71-4dec-802e-e0f9e670ff70", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1845227031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "315a2a3450024ed38e27cd7758f392f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55e46e63-90", "ovs_interfaceid": "55e46e63-9076-400e-a73a-19fc4bb21bec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.564168] env[61987]: DEBUG oslo_vmware.api [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061661, 'name': Destroy_Task} progress is 33%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.654284] env[61987]: DEBUG oslo_concurrency.lockutils [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquiring lock "refresh_cache-80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.654523] env[61987]: DEBUG oslo_concurrency.lockutils [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquired lock "refresh_cache-80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.654771] env[61987]: DEBUG nova.network.neutron [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 804.804582] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061662, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070216} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.810061] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 804.810061] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e7564b-1bbb-45af-83be-2ae70abc6341 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.817572] env[61987]: DEBUG oslo_vmware.api [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061660, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.843789] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] 1f29d5c1-bd01-47dc-9dcf-4e89791120d0/1f29d5c1-bd01-47dc-9dcf-4e89791120d0.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 804.844210] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0185e4c6-e364-496b-9fd9-ab471b8ed338 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.867794] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 804.867794] env[61987]: value = "task-1061666" [ 804.867794] env[61987]: _type = "Task" [ 804.867794] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.878305] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061666, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.970193] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.084s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.972561] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 28.549s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.999681] env[61987]: INFO nova.scheduler.client.report [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Deleted allocations for instance 466a5f5a-f48a-4921-8bcb-44a1840c8141 [ 805.033722] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Releasing lock "refresh_cache-8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.055260] env[61987]: DEBUG oslo_vmware.api [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061661, 'name': Destroy_Task, 'duration_secs': 0.721856} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.055515] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Destroyed the VM [ 805.055871] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Deleting Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 805.056810] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-231769f7-636a-4a0f-b240-e494c38a084b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.066062] env[61987]: DEBUG oslo_vmware.api [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 805.066062] env[61987]: value = "task-1061667" [ 805.066062] env[61987]: _type = "Task" [ 805.066062] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.074850] env[61987]: DEBUG oslo_vmware.api [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061667, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.318300] env[61987]: DEBUG oslo_vmware.api [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061660, 'name': CreateSnapshot_Task, 'duration_secs': 1.18839} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.318723] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Created Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 805.319679] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5075e11c-97d6-4809-8dc2-0a0a0b224b2d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.377859] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061666, 'name': ReconfigVM_Task, 'duration_secs': 0.281167} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.377859] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Reconfigured VM instance instance-0000003f to attach disk [datastore1] 1f29d5c1-bd01-47dc-9dcf-4e89791120d0/1f29d5c1-bd01-47dc-9dcf-4e89791120d0.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 805.378883] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4026200e-a55d-41c6-9da3-87f48c1833f3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.389341] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 805.389341] env[61987]: value = "task-1061668" [ 805.389341] env[61987]: _type = "Task" [ 805.389341] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.397871] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061668, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.479101] env[61987]: DEBUG oslo_concurrency.lockutils [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "8f9aab3e-96ee-44c6-8d25-706a0ff857c4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.479376] env[61987]: DEBUG oslo_concurrency.lockutils [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "8f9aab3e-96ee-44c6-8d25-706a0ff857c4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.510482] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6b576f3a-4e1b-4241-9e17-e492321c7356 tempest-ServersTestManualDisk-1090104942 tempest-ServersTestManualDisk-1090104942-project-member] Lock "466a5f5a-f48a-4921-8bcb-44a1840c8141" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.824s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.559242] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27764ddd-6476-439f-8723-9d45ece90955 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.588694] env[61987]: DEBUG nova.network.neutron [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Updating instance_info_cache with network_info: [{"id": "fc1b4b12-b396-40ac-9eab-a1309f2259e4", "address": "fa:16:3e:ec:35:ab", "network": {"id": "8a9770a4-fed3-4232-ac78-8a1fc2223229", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.118", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "705320a8b6a549d0b197ee1a35404d52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bafe8721-91d4-4127-b215-d9e8e27947dc", "external-id": "nsx-vlan-transportzone-680", "segmentation_id": 680, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfc1b4b12-b3", "ovs_interfaceid": "fc1b4b12-b396-40ac-9eab-a1309f2259e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.592563] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c2b891-4b41-4564-9f2e-3840c9d7e7f9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.601862] env[61987]: DEBUG oslo_vmware.api [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061667, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.605957] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Updating instance '8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2' progress to 83 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 805.842568] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Creating linked-clone VM from snapshot {{(pid=61987) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 805.842789] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3062d6c8-7c9e-4462-a573-3742e63da902 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.851293] env[61987]: DEBUG oslo_vmware.api [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 805.851293] env[61987]: value = "task-1061669" [ 805.851293] env[61987]: _type = "Task" [ 805.851293] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.860266] env[61987]: DEBUG oslo_vmware.api [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061669, 'name': CloneVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.900249] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061668, 'name': Rename_Task, 'duration_secs': 0.13601} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.900521] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 805.900803] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fcfcf283-7dd4-4988-bd66-23dac0bfee88 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.906764] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 805.906764] env[61987]: value = "task-1061670" [ 805.906764] env[61987]: _type = "Task" [ 805.906764] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.915168] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061670, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.991292] env[61987]: DEBUG nova.compute.manager [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 805.995018] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Applying migration context for instance 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2 as it has an incoming, in-progress migration c7ecbdeb-124a-4889-b4e4-2db11acba530. Migration status is post-migrating {{(pid=61987) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 805.996497] env[61987]: INFO nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Updating resource usage from migration c7ecbdeb-124a-4889-b4e4-2db11acba530 [ 806.020728] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 51a36102-795e-47b7-b96a-857e54dc703e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 806.020897] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 4edd0159-6db1-41bd-a616-31a34e984059 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 806.021073] env[61987]: WARNING nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance d038391c-2b32-455c-85d5-68221e7bd065 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 806.021208] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance ad40937f-d5a9-4708-8b81-06087d38d765 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 806.021329] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance f937a5ec-a1d7-41d8-b998-fa18f545b304 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 806.021480] env[61987]: WARNING nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance d566ea53-958b-4a35-989c-771180d95bb4 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 806.021627] env[61987]: WARNING nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 8e68c5a2-6587-45b6-9a76-96b129a0a665 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 806.021845] env[61987]: WARNING nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 1c63bf62-3f70-42b9-b6d7-41336d20f6c8 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 806.021992] env[61987]: WARNING nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 93a27dea-f28a-4cbf-ac37-ff41763ae18d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 806.022128] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 806.022246] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 30c4fff0-c77e-4c11-aced-a040003a6b10 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 806.022396] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Instance with task_state "deleting" is not being actively managed by this compute host but has allocations referencing this compute node (8db744e5-1e14-4e98-9818-695ea4b7dfd0): {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocations during the task state transition. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1780}} [ 806.022955] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 84a4a69b-04db-4546-ba89-9039e382a0c4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 806.022955] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 4fd8fc17-fb13-42e7-9879-9f120fdcb330 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 806.023121] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 1f29d5c1-bd01-47dc-9dcf-4e89791120d0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 806.023492] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Migration c7ecbdeb-124a-4889-b4e4-2db11acba530 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 806.023901] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 806.096955] env[61987]: DEBUG oslo_concurrency.lockutils [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Releasing lock "refresh_cache-80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.097763] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 806.097842] env[61987]: DEBUG oslo_vmware.api [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061667, 'name': RemoveSnapshot_Task, 'duration_secs': 0.637158} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.098014] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d6766406-2590-4ef7-b990-b80352b04a16 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.099868] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Deleted Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 806.100119] env[61987]: INFO nova.compute.manager [None req-d0ed809a-de09-4d3c-8984-0fbdc646676c tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Took 17.17 seconds to snapshot the instance on the hypervisor. [ 806.108498] env[61987]: DEBUG oslo_vmware.api [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for the task: (returnval){ [ 806.108498] env[61987]: value = "task-1061671" [ 806.108498] env[61987]: _type = "Task" [ 806.108498] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.114019] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 806.114336] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ea7b622-7046-45ad-b731-ecc26d4bf50a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.120871] env[61987]: DEBUG oslo_vmware.api [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061671, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.122371] env[61987]: DEBUG oslo_vmware.api [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 806.122371] env[61987]: value = "task-1061672" [ 806.122371] env[61987]: _type = "Task" [ 806.122371] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.131850] env[61987]: DEBUG oslo_vmware.api [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061672, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.362116] env[61987]: DEBUG oslo_vmware.api [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061669, 'name': CloneVM_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.418953] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061670, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.517923] env[61987]: DEBUG oslo_concurrency.lockutils [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.527088] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 59d34772-26a5-4af2-82dc-da932e996745 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 806.618309] env[61987]: DEBUG oslo_vmware.api [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061671, 'name': PowerOffVM_Task, 'duration_secs': 0.18213} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.618593] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 806.619359] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:08:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='3ca61d0c-b87e-4949-982d-be5132df125d',id=36,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-2078042112',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 806.619636] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 806.619802] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 806.620072] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 806.620244] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 806.620469] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 806.620888] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 806.620888] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 806.621043] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 806.621193] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 806.621377] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 806.627286] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0c5c60b-b1af-4683-ac32-12d26712e38a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.649493] env[61987]: DEBUG oslo_vmware.api [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061672, 'name': PowerOnVM_Task, 'duration_secs': 0.506154} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.650795] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 806.651012] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3db5b147-efed-4741-9f91-ebd8ae677d9d tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Updating instance '8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2' progress to 100 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 806.655395] env[61987]: DEBUG oslo_vmware.api [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for the task: (returnval){ [ 806.655395] env[61987]: value = "task-1061674" [ 806.655395] env[61987]: _type = "Task" [ 806.655395] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.665695] env[61987]: DEBUG oslo_vmware.api [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061674, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.862695] env[61987]: DEBUG oslo_vmware.api [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061669, 'name': CloneVM_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.918465] env[61987]: DEBUG oslo_vmware.api [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061670, 'name': PowerOnVM_Task, 'duration_secs': 0.653004} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.918877] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 806.919282] env[61987]: INFO nova.compute.manager [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Took 9.31 seconds to spawn the instance on the hypervisor. [ 806.919688] env[61987]: DEBUG nova.compute.manager [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 806.920786] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-322f601d-eb2a-4b5a-96cf-ffe0fd9674ae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.030628] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance fe22fc1b-61fb-4088-b5fe-b84eb4ec4099 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 807.170715] env[61987]: DEBUG oslo_vmware.api [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061674, 'name': ReconfigVM_Task, 'duration_secs': 0.208279} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.171721] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7203886d-282a-4978-b024-dc2eae87b26c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.195455] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:08:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='3ca61d0c-b87e-4949-982d-be5132df125d',id=36,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-2078042112',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 807.195997] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 807.196377] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 807.196547] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 807.196610] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 807.196775] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 807.197043] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 807.197267] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 807.197498] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 807.197703] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 807.198134] env[61987]: DEBUG nova.virt.hardware [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 807.199887] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59f14b67-0064-4bb5-8081-7accc2963a6a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.207531] env[61987]: DEBUG oslo_vmware.api [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for the task: (returnval){ [ 807.207531] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52be7099-2107-0592-c62c-6d29a90b9f96" [ 807.207531] env[61987]: _type = "Task" [ 807.207531] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.217283] env[61987]: DEBUG oslo_vmware.api [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52be7099-2107-0592-c62c-6d29a90b9f96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.362700] env[61987]: DEBUG oslo_vmware.api [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061669, 'name': CloneVM_Task} progress is 95%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.443111] env[61987]: INFO nova.compute.manager [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Took 45.01 seconds to build instance. [ 807.534957] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 93706a54-1fc4-4aed-8807-ec2c3d3694b4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 807.720950] env[61987]: DEBUG oslo_vmware.api [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52be7099-2107-0592-c62c-6d29a90b9f96, 'name': SearchDatastore_Task, 'duration_secs': 0.008851} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.730261] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Reconfiguring VM instance instance-0000002b to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 807.730261] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b1c3b18a-92b6-49d5-a577-1e28153723c6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.748616] env[61987]: DEBUG oslo_vmware.api [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for the task: (returnval){ [ 807.748616] env[61987]: value = "task-1061675" [ 807.748616] env[61987]: _type = "Task" [ 807.748616] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.759518] env[61987]: DEBUG oslo_vmware.api [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061675, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.862999] env[61987]: DEBUG oslo_vmware.api [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061669, 'name': CloneVM_Task, 'duration_secs': 1.995173} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.863377] env[61987]: INFO nova.virt.vmwareapi.vmops [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Created linked-clone VM from snapshot [ 807.864194] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6c0b781-b076-4ad2-b3c7-f3ae1764e05a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.871704] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Uploading image e8714439-c4a1-44cc-a708-a978ceb46d52 {{(pid=61987) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 807.889449] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Destroying the VM {{(pid=61987) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 807.889449] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c3205a23-2ee8-41b9-91b8-f17829db5280 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.896339] env[61987]: DEBUG oslo_vmware.api [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 807.896339] env[61987]: value = "task-1061676" [ 807.896339] env[61987]: _type = "Task" [ 807.896339] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.905391] env[61987]: DEBUG oslo_vmware.api [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061676, 'name': Destroy_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.945822] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a434a3c7-ea67-4751-ae9a-a17e87563198 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.849s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.041098] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 98d895ee-1d1f-49cd-a055-fe6c8462a692 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 808.260060] env[61987]: DEBUG oslo_vmware.api [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061675, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.321744] env[61987]: DEBUG nova.compute.manager [req-8bdce327-0e01-4a02-bb80-f49b44ab7a12 req-1837de6d-0c81-4f6a-941b-8c5564d598a2 service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Received event network-changed-e7d811c0-d77e-448d-86fe-acdb61c04284 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 808.323027] env[61987]: DEBUG nova.compute.manager [req-8bdce327-0e01-4a02-bb80-f49b44ab7a12 req-1837de6d-0c81-4f6a-941b-8c5564d598a2 service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Refreshing instance network info cache due to event network-changed-e7d811c0-d77e-448d-86fe-acdb61c04284. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 808.323027] env[61987]: DEBUG oslo_concurrency.lockutils [req-8bdce327-0e01-4a02-bb80-f49b44ab7a12 req-1837de6d-0c81-4f6a-941b-8c5564d598a2 service nova] Acquiring lock "refresh_cache-1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.323027] env[61987]: DEBUG oslo_concurrency.lockutils [req-8bdce327-0e01-4a02-bb80-f49b44ab7a12 req-1837de6d-0c81-4f6a-941b-8c5564d598a2 service nova] Acquired lock "refresh_cache-1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.323027] env[61987]: DEBUG nova.network.neutron [req-8bdce327-0e01-4a02-bb80-f49b44ab7a12 req-1837de6d-0c81-4f6a-941b-8c5564d598a2 service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Refreshing network info cache for port e7d811c0-d77e-448d-86fe-acdb61c04284 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 808.407170] env[61987]: DEBUG oslo_vmware.api [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061676, 'name': Destroy_Task, 'duration_secs': 0.358278} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.407170] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Destroyed the VM [ 808.407170] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Deleting Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 808.407300] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-8bbfc4bd-42c7-4a67-95a9-3f3ec450cc38 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.413503] env[61987]: DEBUG oslo_vmware.api [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 808.413503] env[61987]: value = "task-1061677" [ 808.413503] env[61987]: _type = "Task" [ 808.413503] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.421174] env[61987]: DEBUG oslo_vmware.api [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061677, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.547988] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 6ed3f0e9-3c39-4406-9863-931a117961a8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 808.761730] env[61987]: DEBUG oslo_vmware.api [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061675, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.786059] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Volume attach. Driver type: vmdk {{(pid=61987) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 808.786350] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234335', 'volume_id': '41cab2ad-6048-4d47-a5db-77ff38f4f0f0', 'name': 'volume-41cab2ad-6048-4d47-a5db-77ff38f4f0f0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f937a5ec-a1d7-41d8-b998-fa18f545b304', 'attached_at': '', 'detached_at': '', 'volume_id': '41cab2ad-6048-4d47-a5db-77ff38f4f0f0', 'serial': '41cab2ad-6048-4d47-a5db-77ff38f4f0f0'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 808.787814] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb22adc4-d2d4-4100-af36-bc48a74c304f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.804780] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2aea4fbc-31b8-47e9-9a59-5dce72721003 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.833693] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] volume-41cab2ad-6048-4d47-a5db-77ff38f4f0f0/volume-41cab2ad-6048-4d47-a5db-77ff38f4f0f0.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 808.836583] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71dcfab2-c26d-4839-bb7e-8b0f4b9e28f1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.857337] env[61987]: DEBUG oslo_vmware.api [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 808.857337] env[61987]: value = "task-1061678" [ 808.857337] env[61987]: _type = "Task" [ 808.857337] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.869191] env[61987]: DEBUG oslo_vmware.api [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061678, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.922942] env[61987]: DEBUG oslo_vmware.api [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061677, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.027590] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0522c70a-4abf-4f98-b1b3-b522ef661f68 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.028008] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0522c70a-4abf-4f98-b1b3-b522ef661f68 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.028299] env[61987]: DEBUG nova.compute.manager [None req-0522c70a-4abf-4f98-b1b3-b522ef661f68 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Going to confirm migration 2 {{(pid=61987) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5249}} [ 809.053069] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 056c181c-37da-49ac-ae1f-f7f2032c579a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 809.182709] env[61987]: DEBUG nova.network.neutron [req-8bdce327-0e01-4a02-bb80-f49b44ab7a12 req-1837de6d-0c81-4f6a-941b-8c5564d598a2 service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Updated VIF entry in instance network info cache for port e7d811c0-d77e-448d-86fe-acdb61c04284. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 809.183220] env[61987]: DEBUG nova.network.neutron [req-8bdce327-0e01-4a02-bb80-f49b44ab7a12 req-1837de6d-0c81-4f6a-941b-8c5564d598a2 service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Updating instance_info_cache with network_info: [{"id": "e7d811c0-d77e-448d-86fe-acdb61c04284", "address": "fa:16:3e:b0:6d:8c", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7d811c0-d7", "ovs_interfaceid": "e7d811c0-d77e-448d-86fe-acdb61c04284", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.260635] env[61987]: DEBUG oslo_vmware.api [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061675, 'name': ReconfigVM_Task, 'duration_secs': 1.264848} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.260997] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Reconfigured VM instance instance-0000002b to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 809.261926] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef322766-460a-4ae7-b55c-3764f4c1e483 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.286922] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Reconfiguring VM instance instance-0000002b to attach disk [datastore1] 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6/80f13fca-2df1-4aa0-96f1-b61a60c4fdc6.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 809.287386] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c736c3e4-6620-449f-8028-6827a250ec42 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.308319] env[61987]: DEBUG oslo_vmware.api [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for the task: (returnval){ [ 809.308319] env[61987]: value = "task-1061679" [ 809.308319] env[61987]: _type = "Task" [ 809.308319] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.321113] env[61987]: DEBUG oslo_vmware.api [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061679, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.370283] env[61987]: DEBUG oslo_vmware.api [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061678, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.423138] env[61987]: DEBUG oslo_vmware.api [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061677, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.554640] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 1961c2c7-4400-4559-9f7c-bec8e8e263b0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 809.584635] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0522c70a-4abf-4f98-b1b3-b522ef661f68 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "refresh_cache-8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.584859] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0522c70a-4abf-4f98-b1b3-b522ef661f68 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired lock "refresh_cache-8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.585051] env[61987]: DEBUG nova.network.neutron [None req-0522c70a-4abf-4f98-b1b3-b522ef661f68 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 809.585267] env[61987]: DEBUG nova.objects.instance [None req-0522c70a-4abf-4f98-b1b3-b522ef661f68 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lazy-loading 'info_cache' on Instance uuid 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 809.686603] env[61987]: DEBUG oslo_concurrency.lockutils [req-8bdce327-0e01-4a02-bb80-f49b44ab7a12 req-1837de6d-0c81-4f6a-941b-8c5564d598a2 service nova] Releasing lock "refresh_cache-1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.818639] env[61987]: DEBUG oslo_vmware.api [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061679, 'name': ReconfigVM_Task, 'duration_secs': 0.432121} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.818948] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Reconfigured VM instance instance-0000002b to attach disk [datastore1] 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6/80f13fca-2df1-4aa0-96f1-b61a60c4fdc6.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 809.820037] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d50c3da3-eb36-43a4-8f8e-6230aac913e5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.839208] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab5d9cc0-1875-4d3f-94e4-ab425060baea {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.858899] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e789ad1a-c044-4072-abe0-254df14ef5c1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.883317] env[61987]: DEBUG oslo_vmware.api [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061678, 'name': ReconfigVM_Task, 'duration_secs': 0.583915} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.883758] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Reconfigured VM instance instance-00000036 to attach disk [datastore2] volume-41cab2ad-6048-4d47-a5db-77ff38f4f0f0/volume-41cab2ad-6048-4d47-a5db-77ff38f4f0f0.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 809.888857] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e8b6d07-6819-4ba4-8adc-83183804f0a1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.891537] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29a84d7f-7373-453f-81a7-77b83b7683ec {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.908184] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 809.908184] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-863bf8e9-54b5-4aa9-8ce3-43d3b35a9e38 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.909518] env[61987]: DEBUG oslo_vmware.api [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 809.909518] env[61987]: value = "task-1061680" [ 809.909518] env[61987]: _type = "Task" [ 809.909518] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.913832] env[61987]: DEBUG oslo_vmware.api [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for the task: (returnval){ [ 809.913832] env[61987]: value = "task-1061681" [ 809.913832] env[61987]: _type = "Task" [ 809.913832] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.920584] env[61987]: DEBUG oslo_vmware.api [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061680, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.930293] env[61987]: DEBUG oslo_vmware.api [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061681, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.933455] env[61987]: DEBUG oslo_vmware.api [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061677, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.057704] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 8f9aab3e-96ee-44c6-8d25-706a0ff857c4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 810.058080] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 810.059991] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2688MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 810.374564] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d82cced5-cb64-401f-8800-bcb949fe8248 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.381992] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f004c72e-eb62-477b-9473-1e670939a1ed {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.414685] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b7cd033-1811-4009-a87c-a3330bb74f0f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.423687] env[61987]: DEBUG oslo_vmware.api [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061680, 'name': ReconfigVM_Task, 'duration_secs': 0.188807} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.428596] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234335', 'volume_id': '41cab2ad-6048-4d47-a5db-77ff38f4f0f0', 'name': 'volume-41cab2ad-6048-4d47-a5db-77ff38f4f0f0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f937a5ec-a1d7-41d8-b998-fa18f545b304', 'attached_at': '', 'detached_at': '', 'volume_id': '41cab2ad-6048-4d47-a5db-77ff38f4f0f0', 'serial': '41cab2ad-6048-4d47-a5db-77ff38f4f0f0'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 810.433933] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2e4f337-287c-44f6-85fd-998de2bb07d3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.438568] env[61987]: DEBUG oslo_vmware.api [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061681, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.444108] env[61987]: DEBUG oslo_vmware.api [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061677, 'name': RemoveSnapshot_Task, 'duration_secs': 1.799425} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.450259] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Deleted Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 810.452887] env[61987]: DEBUG nova.compute.provider_tree [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.791673] env[61987]: DEBUG nova.network.neutron [None req-0522c70a-4abf-4f98-b1b3-b522ef661f68 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Updating instance_info_cache with network_info: [{"id": "55e46e63-9076-400e-a73a-19fc4bb21bec", "address": "fa:16:3e:09:61:0c", "network": {"id": "59d8d7d6-fd71-4dec-802e-e0f9e670ff70", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1845227031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "315a2a3450024ed38e27cd7758f392f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap55e46e63-90", "ovs_interfaceid": "55e46e63-9076-400e-a73a-19fc4bb21bec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.928466] env[61987]: DEBUG oslo_vmware.api [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061681, 'name': PowerOnVM_Task, 'duration_secs': 0.695691} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.928777] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 810.955388] env[61987]: WARNING nova.compute.manager [None req-4e2b1646-0cce-4791-9601-6f28fcf32dea tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Image not found during snapshot: nova.exception.ImageNotFound: Image e8714439-c4a1-44cc-a708-a978ceb46d52 could not be found. [ 810.957196] env[61987]: DEBUG nova.scheduler.client.report [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 811.295073] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0522c70a-4abf-4f98-b1b3-b522ef661f68 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Releasing lock "refresh_cache-8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.296074] env[61987]: DEBUG nova.objects.instance [None req-0522c70a-4abf-4f98-b1b3-b522ef661f68 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lazy-loading 'migration_context' on Instance uuid 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 811.461654] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61987) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 811.461887] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.489s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.462255] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.429s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.463701] env[61987]: INFO nova.compute.claims [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 811.466427] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 811.466524] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Cleaning up deleted instances {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11701}} [ 811.476314] env[61987]: DEBUG nova.objects.instance [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lazy-loading 'flavor' on Instance uuid f937a5ec-a1d7-41d8-b998-fa18f545b304 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 811.800048] env[61987]: DEBUG nova.objects.base [None req-0522c70a-4abf-4f98-b1b3-b522ef661f68 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Object Instance<8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2> lazy-loaded attributes: info_cache,migration_context {{(pid=61987) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 811.800997] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600c5a87-057a-45b8-b2ac-4097f6ff3f55 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.826879] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22ab6b12-c35d-450e-ae24-ab96991798e2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.832932] env[61987]: DEBUG oslo_vmware.api [None req-0522c70a-4abf-4f98-b1b3-b522ef661f68 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 811.832932] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]521e0ada-4131-0d20-40f5-6fd9eae0c126" [ 811.832932] env[61987]: _type = "Task" [ 811.832932] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.841933] env[61987]: DEBUG oslo_vmware.api [None req-0522c70a-4abf-4f98-b1b3-b522ef661f68 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]521e0ada-4131-0d20-40f5-6fd9eae0c126, 'name': SearchDatastore_Task, 'duration_secs': 0.006996} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.842205] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0522c70a-4abf-4f98-b1b3-b522ef661f68 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.939688] env[61987]: INFO nova.compute.manager [None req-06846ac2-78e6-4418-9aad-c112c506d839 tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Updating instance to original state: 'active' [ 811.947398] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "4fd8fc17-fb13-42e7-9879-9f120fdcb330" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.947632] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "4fd8fc17-fb13-42e7-9879-9f120fdcb330" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.947863] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "4fd8fc17-fb13-42e7-9879-9f120fdcb330-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.948034] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "4fd8fc17-fb13-42e7-9879-9f120fdcb330-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.948241] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "4fd8fc17-fb13-42e7-9879-9f120fdcb330-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.950202] env[61987]: INFO nova.compute.manager [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Terminating instance [ 811.975022] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] There are 18 instances to clean {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11710}} [ 811.975331] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: bb7243dc-a950-4d3e-9b17-944e57cdd56f] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 811.980950] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6f073b0-6a0b-4093-b8b1-5d740b54aff3 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "f937a5ec-a1d7-41d8-b998-fa18f545b304" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.821s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.454140] env[61987]: DEBUG nova.compute.manager [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 812.454471] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 812.455384] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba4b2747-cf78-4ef7-bb2e-91fe015d65fa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.466912] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 812.467236] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c5b0bf9a-9e92-4830-8c6e-e95fe5693999 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.476192] env[61987]: DEBUG oslo_vmware.api [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 812.476192] env[61987]: value = "task-1061682" [ 812.476192] env[61987]: _type = "Task" [ 812.476192] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.482769] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 1afcffda-0bd2-4cbe-8cfb-12a91bb50975] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 812.490124] env[61987]: DEBUG oslo_vmware.api [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061682, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.860174] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96481116-d2d1-47a7-9f9b-9a728fa4f520 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.869852] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-579dc258-f428-4030-84e4-76036f9f5d48 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.904035] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d528a8d8-49bd-4969-af14-871e9821d754 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.911255] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81c6574a-fcdb-46a6-a8dd-47dc772343aa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.926126] env[61987]: DEBUG nova.compute.provider_tree [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 812.986745] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 325183b9-e35b-4f9c-a271-d0fdb05288bb] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 812.988583] env[61987]: DEBUG oslo_vmware.api [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061682, 'name': PowerOffVM_Task, 'duration_secs': 0.187195} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.989046] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 812.989237] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 812.989697] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-37111da4-5472-4bca-8ecc-6caaaafa78d2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.053789] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 813.054169] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 813.054389] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Deleting the datastore file [datastore1] 4fd8fc17-fb13-42e7-9879-9f120fdcb330 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 813.054664] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f73bf4ca-5335-47d0-aaf9-5f8978a3c665 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.060483] env[61987]: DEBUG oslo_vmware.api [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 813.060483] env[61987]: value = "task-1061684" [ 813.060483] env[61987]: _type = "Task" [ 813.060483] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.069176] env[61987]: DEBUG oslo_vmware.api [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061684, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.100147] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquiring lock "80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.101277] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.101607] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquiring lock "80f13fca-2df1-4aa0-96f1-b61a60c4fdc6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.101898] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "80f13fca-2df1-4aa0-96f1-b61a60c4fdc6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.102170] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "80f13fca-2df1-4aa0-96f1-b61a60c4fdc6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.104445] env[61987]: INFO nova.compute.manager [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Terminating instance [ 813.429886] env[61987]: DEBUG nova.scheduler.client.report [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 813.489798] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: d55395cc-0b92-4e99-9b80-913c19f6c7f4] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 813.572078] env[61987]: DEBUG oslo_vmware.api [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061684, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128516} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.572395] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 813.573050] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 813.573050] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 813.573050] env[61987]: INFO nova.compute.manager [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Took 1.12 seconds to destroy the instance on the hypervisor. [ 813.573267] env[61987]: DEBUG oslo.service.loopingcall [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 813.573415] env[61987]: DEBUG nova.compute.manager [-] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 813.573624] env[61987]: DEBUG nova.network.neutron [-] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 813.611423] env[61987]: DEBUG nova.compute.manager [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 813.611665] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 813.612754] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c02af62-2285-41e8-a0a5-101dfb97ffd8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.620420] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 813.620714] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-97b4fd7f-ffcc-49aa-af50-2e2984051127 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.626758] env[61987]: DEBUG oslo_vmware.api [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for the task: (returnval){ [ 813.626758] env[61987]: value = "task-1061685" [ 813.626758] env[61987]: _type = "Task" [ 813.626758] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.635754] env[61987]: DEBUG oslo_vmware.api [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061685, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.888881] env[61987]: DEBUG nova.compute.manager [req-d1b4ad35-da0c-4b06-9f2c-19de3a4f3a4f req-faae299a-a0b6-48d4-ae03-d7d0b58d99c7 service nova] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Received event network-vif-deleted-a3d8a0f6-c3dc-42d9-b4e1-903a299b601e {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 813.889609] env[61987]: INFO nova.compute.manager [req-d1b4ad35-da0c-4b06-9f2c-19de3a4f3a4f req-faae299a-a0b6-48d4-ae03-d7d0b58d99c7 service nova] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Neutron deleted interface a3d8a0f6-c3dc-42d9-b4e1-903a299b601e; detaching it from the instance and deleting it from the info cache [ 813.889845] env[61987]: DEBUG nova.network.neutron [req-d1b4ad35-da0c-4b06-9f2c-19de3a4f3a4f req-faae299a-a0b6-48d4-ae03-d7d0b58d99c7 service nova] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.908457] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "742297eb-c45d-4f45-85d2-e9bdd3106d1e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.908712] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "742297eb-c45d-4f45-85d2-e9bdd3106d1e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.934608] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.472s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.935202] env[61987]: DEBUG nova.compute.manager [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 813.937815] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.571s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.939484] env[61987]: INFO nova.compute.claims [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 813.993365] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 93174a10-f8b1-4789-ab3c-dda07bdddaa8] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 814.136862] env[61987]: DEBUG oslo_vmware.api [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061685, 'name': PowerOffVM_Task, 'duration_secs': 0.262362} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.137197] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 814.137377] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 814.137638] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a028730d-c5e5-4d52-83f6-438a64f95ee8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.216276] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 814.216496] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 814.216659] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Deleting the datastore file [datastore1] 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 814.216947] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c4c1c119-e9fd-467f-b395-c7fd798b1b4b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.222902] env[61987]: DEBUG oslo_vmware.api [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for the task: (returnval){ [ 814.222902] env[61987]: value = "task-1061687" [ 814.222902] env[61987]: _type = "Task" [ 814.222902] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.230838] env[61987]: DEBUG oslo_vmware.api [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061687, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.305046] env[61987]: DEBUG nova.network.neutron [-] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.393194] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-879be08e-868c-4913-bf19-5d107e1563b3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.403072] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-604921de-8d0c-480b-a468-2443b4443c47 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.414428] env[61987]: DEBUG nova.compute.manager [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 814.435684] env[61987]: DEBUG nova.compute.manager [req-d1b4ad35-da0c-4b06-9f2c-19de3a4f3a4f req-faae299a-a0b6-48d4-ae03-d7d0b58d99c7 service nova] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Detach interface failed, port_id=a3d8a0f6-c3dc-42d9-b4e1-903a299b601e, reason: Instance 4fd8fc17-fb13-42e7-9879-9f120fdcb330 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 814.444403] env[61987]: DEBUG nova.compute.utils [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 814.447991] env[61987]: DEBUG nova.compute.manager [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Not allocating networking since 'none' was specified. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 814.496182] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: e9f0b620-b0af-4fe2-ac5d-e3f6c8c998b1] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 814.733064] env[61987]: DEBUG oslo_vmware.api [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Task: {'id': task-1061687, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164405} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.733301] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 814.733489] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 814.733682] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 814.733865] env[61987]: INFO nova.compute.manager [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Took 1.12 seconds to destroy the instance on the hypervisor. [ 814.734164] env[61987]: DEBUG oslo.service.loopingcall [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 814.734381] env[61987]: DEBUG nova.compute.manager [-] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 814.734479] env[61987]: DEBUG nova.network.neutron [-] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 814.808018] env[61987]: INFO nova.compute.manager [-] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Took 1.23 seconds to deallocate network for instance. [ 814.935603] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.948882] env[61987]: DEBUG nova.compute.manager [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 815.000186] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 8ff4ebb9-57ea-49d2-8d3c-5876da61e77b] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 815.250616] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab9eeb9-9c2a-4a36-a267-9d5dc3c3bec7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.258195] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b68f63b0-b363-40cf-8ee6-a5f5cc479985 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.289113] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5365dc4c-35ea-4bf3-ab32-c0cbec51c140 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.296705] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dbd2770-c477-4e6e-bd08-cc098433674f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.309908] env[61987]: DEBUG nova.compute.provider_tree [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.314344] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.458497] env[61987]: DEBUG nova.network.neutron [-] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.503515] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 49e0d752-6f88-45c4-bd6a-d9fe4f526a2f] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 815.813036] env[61987]: DEBUG nova.scheduler.client.report [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 815.918154] env[61987]: DEBUG nova.compute.manager [req-b82b3753-481b-425c-9076-dcba171764bc req-76665e2a-2ec4-4555-99db-4272def6ede5 service nova] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Received event network-vif-deleted-fc1b4b12-b396-40ac-9eab-a1309f2259e4 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 815.961699] env[61987]: INFO nova.compute.manager [-] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Took 1.23 seconds to deallocate network for instance. [ 815.962813] env[61987]: DEBUG nova.compute.manager [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 815.993694] env[61987]: DEBUG nova.virt.hardware [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 815.994193] env[61987]: DEBUG nova.virt.hardware [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 815.994554] env[61987]: DEBUG nova.virt.hardware [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 815.994951] env[61987]: DEBUG nova.virt.hardware [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 815.995279] env[61987]: DEBUG nova.virt.hardware [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 815.995606] env[61987]: DEBUG nova.virt.hardware [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 815.995959] env[61987]: DEBUG nova.virt.hardware [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 815.996279] env[61987]: DEBUG nova.virt.hardware [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 815.996601] env[61987]: DEBUG nova.virt.hardware [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 815.996942] env[61987]: DEBUG nova.virt.hardware [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 815.997293] env[61987]: DEBUG nova.virt.hardware [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 815.998716] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-093e35e8-434e-4bd2-b91c-e439041a8981 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.009572] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: fa943137-caf1-4b41-b644-c9146f72f956] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 816.014259] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20bc5070-54b4-423e-8dd4-df95572930e9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.035240] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Instance VIF info [] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 816.040919] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Creating folder: Project (6250c7d70eb546a7bcdfde978116221a). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 816.041609] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d2d61ca1-0656-45ac-8477-7e8d496fbd9b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.051843] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Created folder: Project (6250c7d70eb546a7bcdfde978116221a) in parent group-v234219. [ 816.052163] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Creating folder: Instances. Parent ref: group-v234338. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 816.052380] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d144be4b-d327-4dd8-aa1d-329c1b23792f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.062134] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Created folder: Instances in parent group-v234338. [ 816.062387] env[61987]: DEBUG oslo.service.loopingcall [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.062582] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 816.062788] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3af5691a-41a8-46b3-9cf4-38f924b6157b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.079165] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 816.079165] env[61987]: value = "task-1061690" [ 816.079165] env[61987]: _type = "Task" [ 816.079165] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.086800] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061690, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.317592] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.318225] env[61987]: DEBUG nova.compute.manager [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 816.321024] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.700s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.321236] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.323641] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.592s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.323857] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.325618] env[61987]: DEBUG oslo_concurrency.lockutils [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 36.314s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.327238] env[61987]: INFO nova.compute.claims [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 816.350976] env[61987]: INFO nova.scheduler.client.report [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Deleted allocations for instance 8e68c5a2-6587-45b6-9a76-96b129a0a665 [ 816.354237] env[61987]: INFO nova.scheduler.client.report [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Deleted allocations for instance d566ea53-958b-4a35-989c-771180d95bb4 [ 816.471287] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.525061] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 0b4b3bdb-7393-47ab-abc1-44c47496cb3a] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 816.589515] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061690, 'name': CreateVM_Task, 'duration_secs': 0.311286} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.589704] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 816.590336] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.590443] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.590723] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 816.590981] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-999dcce7-6ed9-4269-bc47-d0fda854244f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.596457] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for the task: (returnval){ [ 816.596457] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52841918-149c-9952-ba94-2bb00a4c9884" [ 816.596457] env[61987]: _type = "Task" [ 816.596457] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.603441] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52841918-149c-9952-ba94-2bb00a4c9884, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.834047] env[61987]: DEBUG nova.compute.utils [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 816.841132] env[61987]: DEBUG nova.compute.manager [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 816.841132] env[61987]: DEBUG nova.network.neutron [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 816.865087] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6a733f6f-ad35-4f72-84d6-0d4194ea80ac tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "8e68c5a2-6587-45b6-9a76-96b129a0a665" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.744s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.868260] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0efb9951-e057-401a-8b5f-00e35ba7e178 tempest-MultipleCreateTestJSON-478176008 tempest-MultipleCreateTestJSON-478176008-project-member] Lock "d566ea53-958b-4a35-989c-771180d95bb4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.935s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.891564] env[61987]: DEBUG nova.policy [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'da739850d04b4251a862f12d6ffe258c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e802bb0ab61a41a79664c7e12fa0e2bf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 817.028771] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 02fe490b-75c8-406b-98e0-aa200f249673] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 817.107708] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52841918-149c-9952-ba94-2bb00a4c9884, 'name': SearchDatastore_Task, 'duration_secs': 0.00986} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.108034] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.108281] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 817.108515] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.108658] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.108836] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 817.109144] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-58e6d016-3f64-4d91-8a4c-71652d67fc9e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.118054] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 817.118242] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 817.119000] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c43360d-b11d-4e16-8e96-ad38bdcced0b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.126547] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for the task: (returnval){ [ 817.126547] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]527123c9-da2b-35c9-dd8d-49fc54afcbb1" [ 817.126547] env[61987]: _type = "Task" [ 817.126547] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.134502] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]527123c9-da2b-35c9-dd8d-49fc54afcbb1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.222167] env[61987]: DEBUG nova.network.neutron [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Successfully created port: b4c377b2-8e62-4cde-ac10-05c4cd2bde7d {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 817.340863] env[61987]: DEBUG nova.compute.manager [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 817.531438] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: ea6c2120-a69a-42c7-bc92-3e74539c008d] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 817.644488] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]527123c9-da2b-35c9-dd8d-49fc54afcbb1, 'name': SearchDatastore_Task, 'duration_secs': 0.008702} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.645544] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e740d16b-79ce-46cd-a192-e2999412a85c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.653044] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for the task: (returnval){ [ 817.653044] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5279b5af-aac8-f765-790b-ca7787b54a6b" [ 817.653044] env[61987]: _type = "Task" [ 817.653044] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.665429] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5279b5af-aac8-f765-790b-ca7787b54a6b, 'name': SearchDatastore_Task, 'duration_secs': 0.008869} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.665533] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.665734] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 59d34772-26a5-4af2-82dc-da932e996745/59d34772-26a5-4af2-82dc-da932e996745.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 817.665997] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6e06bfef-34cc-4acd-b1b4-028a093d5f05 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.671967] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for the task: (returnval){ [ 817.671967] env[61987]: value = "task-1061691" [ 817.671967] env[61987]: _type = "Task" [ 817.671967] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.673451] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb66110-ce43-4284-8586-370358318d8d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.686505] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22ae1e24-37d5-46d6-ac51-6d83cccee264 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.689727] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061691, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.716592] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd4f3c5-7c69-4dc6-8ba4-570cc2a37081 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.724176] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9912d86f-21c4-43ee-a16e-e62a86f2d171 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.739414] env[61987]: DEBUG nova.compute.provider_tree [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.035535] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 466a5f5a-f48a-4921-8bcb-44a1840c8141] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 818.187850] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061691, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.242909] env[61987]: DEBUG nova.scheduler.client.report [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 818.352872] env[61987]: DEBUG nova.compute.manager [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 818.379572] env[61987]: DEBUG nova.virt.hardware [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 818.379853] env[61987]: DEBUG nova.virt.hardware [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 818.380075] env[61987]: DEBUG nova.virt.hardware [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 818.380365] env[61987]: DEBUG nova.virt.hardware [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 818.380587] env[61987]: DEBUG nova.virt.hardware [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 818.380794] env[61987]: DEBUG nova.virt.hardware [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 818.381124] env[61987]: DEBUG nova.virt.hardware [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 818.381361] env[61987]: DEBUG nova.virt.hardware [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 818.381608] env[61987]: DEBUG nova.virt.hardware [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 818.381808] env[61987]: DEBUG nova.virt.hardware [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 818.382048] env[61987]: DEBUG nova.virt.hardware [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 818.382977] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2101d7-f53c-4af1-8b16-42e2c4b7ae29 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.391791] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1013335-53e3-49c1-a355-8575172e2218 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.538776] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 630e0959-6d7f-401d-b84c-1c098f54bf60] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 818.662319] env[61987]: DEBUG nova.compute.manager [req-0bcf2bd2-fe80-454f-bf86-160bdeead712 req-9b3ff9b4-4ac4-4d03-b861-2b26dc5450a3 service nova] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Received event network-vif-plugged-b4c377b2-8e62-4cde-ac10-05c4cd2bde7d {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 818.662319] env[61987]: DEBUG oslo_concurrency.lockutils [req-0bcf2bd2-fe80-454f-bf86-160bdeead712 req-9b3ff9b4-4ac4-4d03-b861-2b26dc5450a3 service nova] Acquiring lock "fe22fc1b-61fb-4088-b5fe-b84eb4ec4099-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.662526] env[61987]: DEBUG oslo_concurrency.lockutils [req-0bcf2bd2-fe80-454f-bf86-160bdeead712 req-9b3ff9b4-4ac4-4d03-b861-2b26dc5450a3 service nova] Lock "fe22fc1b-61fb-4088-b5fe-b84eb4ec4099-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.662743] env[61987]: DEBUG oslo_concurrency.lockutils [req-0bcf2bd2-fe80-454f-bf86-160bdeead712 req-9b3ff9b4-4ac4-4d03-b861-2b26dc5450a3 service nova] Lock "fe22fc1b-61fb-4088-b5fe-b84eb4ec4099-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.663104] env[61987]: DEBUG nova.compute.manager [req-0bcf2bd2-fe80-454f-bf86-160bdeead712 req-9b3ff9b4-4ac4-4d03-b861-2b26dc5450a3 service nova] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] No waiting events found dispatching network-vif-plugged-b4c377b2-8e62-4cde-ac10-05c4cd2bde7d {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 818.663104] env[61987]: WARNING nova.compute.manager [req-0bcf2bd2-fe80-454f-bf86-160bdeead712 req-9b3ff9b4-4ac4-4d03-b861-2b26dc5450a3 service nova] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Received unexpected event network-vif-plugged-b4c377b2-8e62-4cde-ac10-05c4cd2bde7d for instance with vm_state building and task_state spawning. [ 818.683700] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061691, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.522859} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.683976] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 59d34772-26a5-4af2-82dc-da932e996745/59d34772-26a5-4af2-82dc-da932e996745.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 818.684249] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 818.684525] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e8d718a0-833c-4af3-a1a2-282b6e8bdc13 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.691335] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for the task: (returnval){ [ 818.691335] env[61987]: value = "task-1061692" [ 818.691335] env[61987]: _type = "Task" [ 818.691335] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.700384] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061692, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.751094] env[61987]: DEBUG oslo_concurrency.lockutils [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.757267] env[61987]: DEBUG nova.compute.manager [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 818.757267] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.303s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.761607] env[61987]: INFO nova.compute.claims [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 818.769412] env[61987]: DEBUG nova.network.neutron [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Successfully updated port: b4c377b2-8e62-4cde-ac10-05c4cd2bde7d {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 819.041761] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 89a97f81-bcd9-49e5-bbd3-9140f6638a20] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 819.203431] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061692, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.246504} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.203858] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 819.204733] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d529b2c-e1c1-4643-b9aa-235c804560fb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.224974] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 59d34772-26a5-4af2-82dc-da932e996745/59d34772-26a5-4af2-82dc-da932e996745.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 819.225338] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3aa18c5e-d6cf-4e0b-8b0a-7d60cfd8ef26 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.244877] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for the task: (returnval){ [ 819.244877] env[61987]: value = "task-1061693" [ 819.244877] env[61987]: _type = "Task" [ 819.244877] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.253067] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061693, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.266287] env[61987]: DEBUG nova.compute.utils [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 819.270270] env[61987]: DEBUG nova.compute.manager [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 819.270270] env[61987]: DEBUG nova.network.neutron [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 819.275126] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Acquiring lock "refresh_cache-fe22fc1b-61fb-4088-b5fe-b84eb4ec4099" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.275126] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Acquired lock "refresh_cache-fe22fc1b-61fb-4088-b5fe-b84eb4ec4099" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.275126] env[61987]: DEBUG nova.network.neutron [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 819.332409] env[61987]: DEBUG nova.policy [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '750d9bd6f45c4fc3bd2b23c254fba92f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd32dc833164a428dad56d1eeb9a7c3eb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 819.544844] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 6c0bc598-eae5-4e5b-a3b2-b4bb90bc8d59] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 819.622235] env[61987]: DEBUG nova.network.neutron [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Successfully created port: 6fcb037d-fd48-490d-9898-b952ffaad4ef {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 819.754443] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061693, 'name': ReconfigVM_Task, 'duration_secs': 0.444293} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.754689] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 59d34772-26a5-4af2-82dc-da932e996745/59d34772-26a5-4af2-82dc-da932e996745.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.755354] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-15d4d88b-06ec-4419-91e6-138467dde6d6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.761555] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for the task: (returnval){ [ 819.761555] env[61987]: value = "task-1061694" [ 819.761555] env[61987]: _type = "Task" [ 819.761555] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.770277] env[61987]: DEBUG nova.compute.utils [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 819.771689] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061694, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.812262] env[61987]: DEBUG nova.network.neutron [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 820.012318] env[61987]: DEBUG nova.network.neutron [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Updating instance_info_cache with network_info: [{"id": "b4c377b2-8e62-4cde-ac10-05c4cd2bde7d", "address": "fa:16:3e:6f:ef:57", "network": {"id": "2bc28674-9598-4875-b1bc-4265efac93fe", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1333586968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e802bb0ab61a41a79664c7e12fa0e2bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dad4f433-bb0b-45c7-8040-972ef2277f75", "external-id": "nsx-vlan-transportzone-451", "segmentation_id": 451, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4c377b2-8e", "ovs_interfaceid": "b4c377b2-8e62-4cde-ac10-05c4cd2bde7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.048500] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: e6c56d86-e428-49fb-9944-293ad5e5948c] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 820.133137] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abb67ce7-8d51-4255-a5f3-90fa801e7a4b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.142089] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d38cb2e-dbad-4483-a690-d790101f23fd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.174993] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd3471f-36af-458f-9388-43096c59b2a3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.183148] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef237dd3-1669-46af-aecb-1308262ad313 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.198729] env[61987]: DEBUG nova.compute.provider_tree [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.272094] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061694, 'name': Rename_Task, 'duration_secs': 0.132233} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.272317] env[61987]: DEBUG nova.compute.manager [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 820.274902] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 820.275659] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d14cd488-e865-492a-981a-17854a4fd68e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.282162] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for the task: (returnval){ [ 820.282162] env[61987]: value = "task-1061695" [ 820.282162] env[61987]: _type = "Task" [ 820.282162] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.290210] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061695, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.515048] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Releasing lock "refresh_cache-fe22fc1b-61fb-4088-b5fe-b84eb4ec4099" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.515195] env[61987]: DEBUG nova.compute.manager [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Instance network_info: |[{"id": "b4c377b2-8e62-4cde-ac10-05c4cd2bde7d", "address": "fa:16:3e:6f:ef:57", "network": {"id": "2bc28674-9598-4875-b1bc-4265efac93fe", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1333586968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e802bb0ab61a41a79664c7e12fa0e2bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dad4f433-bb0b-45c7-8040-972ef2277f75", "external-id": "nsx-vlan-transportzone-451", "segmentation_id": 451, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4c377b2-8e", "ovs_interfaceid": "b4c377b2-8e62-4cde-ac10-05c4cd2bde7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 820.515585] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6f:ef:57', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dad4f433-bb0b-45c7-8040-972ef2277f75', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b4c377b2-8e62-4cde-ac10-05c4cd2bde7d', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 820.523350] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Creating folder: Project (e802bb0ab61a41a79664c7e12fa0e2bf). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 820.523701] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d7e4012c-101c-4bf3-9624-385742744b97 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.536292] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Created folder: Project (e802bb0ab61a41a79664c7e12fa0e2bf) in parent group-v234219. [ 820.536502] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Creating folder: Instances. Parent ref: group-v234341. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 820.536773] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-43bdad42-b18d-4876-bab1-084ddbbb0013 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.546189] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Created folder: Instances in parent group-v234341. [ 820.546512] env[61987]: DEBUG oslo.service.loopingcall [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.546646] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 820.546902] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-be499bee-0e5d-4457-bf58-da2803476e04 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.561648] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 6bf0d1f1-d079-4956-bc9a-86ba66885f3a] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 820.568602] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 820.568602] env[61987]: value = "task-1061698" [ 820.568602] env[61987]: _type = "Task" [ 820.568602] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.578107] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061698, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.693620] env[61987]: DEBUG nova.compute.manager [req-0d23d5fb-7df7-43db-b8fe-6ef6018e5b25 req-f24638f4-c515-4bf5-989a-531272fe6c03 service nova] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Received event network-changed-b4c377b2-8e62-4cde-ac10-05c4cd2bde7d {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 820.693817] env[61987]: DEBUG nova.compute.manager [req-0d23d5fb-7df7-43db-b8fe-6ef6018e5b25 req-f24638f4-c515-4bf5-989a-531272fe6c03 service nova] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Refreshing instance network info cache due to event network-changed-b4c377b2-8e62-4cde-ac10-05c4cd2bde7d. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 820.694070] env[61987]: DEBUG oslo_concurrency.lockutils [req-0d23d5fb-7df7-43db-b8fe-6ef6018e5b25 req-f24638f4-c515-4bf5-989a-531272fe6c03 service nova] Acquiring lock "refresh_cache-fe22fc1b-61fb-4088-b5fe-b84eb4ec4099" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.694241] env[61987]: DEBUG oslo_concurrency.lockutils [req-0d23d5fb-7df7-43db-b8fe-6ef6018e5b25 req-f24638f4-c515-4bf5-989a-531272fe6c03 service nova] Acquired lock "refresh_cache-fe22fc1b-61fb-4088-b5fe-b84eb4ec4099" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.694410] env[61987]: DEBUG nova.network.neutron [req-0d23d5fb-7df7-43db-b8fe-6ef6018e5b25 req-f24638f4-c515-4bf5-989a-531272fe6c03 service nova] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Refreshing network info cache for port b4c377b2-8e62-4cde-ac10-05c4cd2bde7d {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 820.702272] env[61987]: DEBUG nova.scheduler.client.report [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 820.792865] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061695, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.065047] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 821.065207] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Cleaning up deleted instances with incomplete migration {{(pid=61987) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11739}} [ 821.078519] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061698, 'name': CreateVM_Task, 'duration_secs': 0.312635} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.078739] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 821.079337] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.079519] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.079838] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 821.080386] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-989a0275-aa84-4f22-8b69-54a159916569 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.085358] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Waiting for the task: (returnval){ [ 821.085358] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52476a3a-9fb9-d122-9879-cff58a8e0955" [ 821.085358] env[61987]: _type = "Task" [ 821.085358] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.092711] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52476a3a-9fb9-d122-9879-cff58a8e0955, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.182761] env[61987]: DEBUG nova.network.neutron [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Successfully updated port: 6fcb037d-fd48-490d-9898-b952ffaad4ef {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 821.207103] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.451s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.207621] env[61987]: DEBUG nova.compute.manager [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 821.210979] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 39.393s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.211206] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.213344] env[61987]: DEBUG oslo_concurrency.lockutils [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 37.557s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.213530] env[61987]: DEBUG oslo_concurrency.lockutils [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.215746] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.063s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.217219] env[61987]: INFO nova.compute.claims [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 821.244779] env[61987]: INFO nova.scheduler.client.report [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Deleted allocations for instance d038391c-2b32-455c-85d5-68221e7bd065 [ 821.253738] env[61987]: INFO nova.scheduler.client.report [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Deleted allocations for instance 1c63bf62-3f70-42b9-b6d7-41336d20f6c8 [ 821.286838] env[61987]: DEBUG nova.compute.manager [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 821.301010] env[61987]: DEBUG oslo_vmware.api [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061695, 'name': PowerOnVM_Task, 'duration_secs': 0.529412} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.301010] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 821.301385] env[61987]: INFO nova.compute.manager [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Took 5.34 seconds to spawn the instance on the hypervisor. [ 821.301737] env[61987]: DEBUG nova.compute.manager [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 821.302647] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc5edefa-bc0f-4ad1-8dfc-fc6715258c40 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.323094] env[61987]: DEBUG nova.virt.hardware [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:09:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=1,extra_specs={hw_rng:allowed='True'},flavorid='2017050287',id=37,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_1-22031768',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 821.323094] env[61987]: DEBUG nova.virt.hardware [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 821.323094] env[61987]: DEBUG nova.virt.hardware [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 821.323094] env[61987]: DEBUG nova.virt.hardware [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 821.323094] env[61987]: DEBUG nova.virt.hardware [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 821.323094] env[61987]: DEBUG nova.virt.hardware [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 821.323094] env[61987]: DEBUG nova.virt.hardware [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 821.323094] env[61987]: DEBUG nova.virt.hardware [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 821.323949] env[61987]: DEBUG nova.virt.hardware [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 821.324307] env[61987]: DEBUG nova.virt.hardware [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 821.324625] env[61987]: DEBUG nova.virt.hardware [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 821.325655] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2993b84d-0ee4-455d-a7a9-3f0bf5171066 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.334629] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5026a177-2766-4b14-a39e-ed51740dd855 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.466773] env[61987]: DEBUG nova.network.neutron [req-0d23d5fb-7df7-43db-b8fe-6ef6018e5b25 req-f24638f4-c515-4bf5-989a-531272fe6c03 service nova] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Updated VIF entry in instance network info cache for port b4c377b2-8e62-4cde-ac10-05c4cd2bde7d. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 821.467270] env[61987]: DEBUG nova.network.neutron [req-0d23d5fb-7df7-43db-b8fe-6ef6018e5b25 req-f24638f4-c515-4bf5-989a-531272fe6c03 service nova] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Updating instance_info_cache with network_info: [{"id": "b4c377b2-8e62-4cde-ac10-05c4cd2bde7d", "address": "fa:16:3e:6f:ef:57", "network": {"id": "2bc28674-9598-4875-b1bc-4265efac93fe", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1333586968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e802bb0ab61a41a79664c7e12fa0e2bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dad4f433-bb0b-45c7-8040-972ef2277f75", "external-id": "nsx-vlan-transportzone-451", "segmentation_id": 451, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4c377b2-8e", "ovs_interfaceid": "b4c377b2-8e62-4cde-ac10-05c4cd2bde7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.568951] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 821.595427] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52476a3a-9fb9-d122-9879-cff58a8e0955, 'name': SearchDatastore_Task, 'duration_secs': 0.01113} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.595803] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.595981] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 821.596522] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.596522] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.596730] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 821.596864] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d7a3a3d2-35de-4b13-956b-dd68898f402c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.606030] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 821.606030] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 821.606613] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04b626d2-7c44-413a-b308-46213dde5ebc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.612158] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Waiting for the task: (returnval){ [ 821.612158] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]522af9d9-b02f-e6b1-3130-321cd23789d7" [ 821.612158] env[61987]: _type = "Task" [ 821.612158] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.619651] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]522af9d9-b02f-e6b1-3130-321cd23789d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.687047] env[61987]: DEBUG oslo_concurrency.lockutils [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquiring lock "refresh_cache-93706a54-1fc4-4aed-8807-ec2c3d3694b4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.687157] env[61987]: DEBUG oslo_concurrency.lockutils [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquired lock "refresh_cache-93706a54-1fc4-4aed-8807-ec2c3d3694b4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.687323] env[61987]: DEBUG nova.network.neutron [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 821.723564] env[61987]: DEBUG nova.compute.utils [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 821.724937] env[61987]: DEBUG nova.compute.manager [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 821.725116] env[61987]: DEBUG nova.network.neutron [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 821.753684] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c32d6369-7efe-4066-b63a-5e9663fcfbb6 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Lock "d038391c-2b32-455c-85d5-68221e7bd065" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 43.920s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.760386] env[61987]: DEBUG oslo_concurrency.lockutils [None req-01fdf338-7878-4f0e-85a3-01479d49851d tempest-InstanceActionsNegativeTestJSON-1877622913 tempest-InstanceActionsNegativeTestJSON-1877622913-project-member] Lock "1c63bf62-3f70-42b9-b6d7-41336d20f6c8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 41.766s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.824125] env[61987]: INFO nova.compute.manager [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Took 44.81 seconds to build instance. [ 821.843888] env[61987]: DEBUG nova.policy [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a585461a1a504c08a3186033e9160eb9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8dffc4d03e514fb8932b80718375084b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 821.970410] env[61987]: DEBUG oslo_concurrency.lockutils [req-0d23d5fb-7df7-43db-b8fe-6ef6018e5b25 req-f24638f4-c515-4bf5-989a-531272fe6c03 service nova] Releasing lock "refresh_cache-fe22fc1b-61fb-4088-b5fe-b84eb4ec4099" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.123227] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]522af9d9-b02f-e6b1-3130-321cd23789d7, 'name': SearchDatastore_Task, 'duration_secs': 0.008254} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.123951] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-919ae6ae-65a9-402c-a634-ca3227fcfda7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.129494] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Waiting for the task: (returnval){ [ 822.129494] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52a0abc8-efa2-3fbc-45a6-0edebda98902" [ 822.129494] env[61987]: _type = "Task" [ 822.129494] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.138321] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52a0abc8-efa2-3fbc-45a6-0edebda98902, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.230450] env[61987]: DEBUG nova.compute.manager [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 822.249896] env[61987]: DEBUG nova.network.neutron [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 822.329153] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c0ff8bf6-c8b5-4aaf-8154-255df458bc3e tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Lock "59d34772-26a5-4af2-82dc-da932e996745" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.323s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.474215] env[61987]: DEBUG nova.network.neutron [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Updating instance_info_cache with network_info: [{"id": "6fcb037d-fd48-490d-9898-b952ffaad4ef", "address": "fa:16:3e:14:a7:3a", "network": {"id": "35c53c0f-af9d-4934-84c7-ea4c5252fa58", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-343092073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d32dc833164a428dad56d1eeb9a7c3eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "790c811b-3fa6-49f8-87ac-c51450911137", "external-id": "nsx-vlan-transportzone-908", "segmentation_id": 908, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fcb037d-fd", "ovs_interfaceid": "6fcb037d-fd48-490d-9898-b952ffaad4ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.573592] env[61987]: DEBUG nova.network.neutron [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Successfully created port: 69ce8bb3-dc28-422e-96a8-833062bd6b77 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 822.593495] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8796b361-1220-44a3-a276-454199613bf3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.602182] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9114117c-de6a-4193-969f-a9c6e56e57f5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.646127] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1859ac7e-0c7b-4d37-9619-aedf2ea3d436 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.658738] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52a0abc8-efa2-3fbc-45a6-0edebda98902, 'name': SearchDatastore_Task, 'duration_secs': 0.009079} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.660993] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.661403] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] fe22fc1b-61fb-4088-b5fe-b84eb4ec4099/fe22fc1b-61fb-4088-b5fe-b84eb4ec4099.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 822.662063] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4f17ffd5-a6ac-437c-8cfc-fa1085d4a538 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.665419] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc1e9752-2ed3-486e-a8ee-d44fcca65269 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.681129] env[61987]: DEBUG nova.compute.provider_tree [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.683504] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Waiting for the task: (returnval){ [ 822.683504] env[61987]: value = "task-1061699" [ 822.683504] env[61987]: _type = "Task" [ 822.683504] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.692332] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': task-1061699, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.928431] env[61987]: DEBUG nova.compute.manager [req-e8671de3-9c14-4b55-b238-468574336d59 req-67efda73-049d-4127-86fb-a80b778e8dbd service nova] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Received event network-vif-plugged-6fcb037d-fd48-490d-9898-b952ffaad4ef {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 822.928542] env[61987]: DEBUG oslo_concurrency.lockutils [req-e8671de3-9c14-4b55-b238-468574336d59 req-67efda73-049d-4127-86fb-a80b778e8dbd service nova] Acquiring lock "93706a54-1fc4-4aed-8807-ec2c3d3694b4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.929237] env[61987]: DEBUG oslo_concurrency.lockutils [req-e8671de3-9c14-4b55-b238-468574336d59 req-67efda73-049d-4127-86fb-a80b778e8dbd service nova] Lock "93706a54-1fc4-4aed-8807-ec2c3d3694b4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.932159] env[61987]: DEBUG oslo_concurrency.lockutils [req-e8671de3-9c14-4b55-b238-468574336d59 req-67efda73-049d-4127-86fb-a80b778e8dbd service nova] Lock "93706a54-1fc4-4aed-8807-ec2c3d3694b4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.932159] env[61987]: DEBUG nova.compute.manager [req-e8671de3-9c14-4b55-b238-468574336d59 req-67efda73-049d-4127-86fb-a80b778e8dbd service nova] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] No waiting events found dispatching network-vif-plugged-6fcb037d-fd48-490d-9898-b952ffaad4ef {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 822.932159] env[61987]: WARNING nova.compute.manager [req-e8671de3-9c14-4b55-b238-468574336d59 req-67efda73-049d-4127-86fb-a80b778e8dbd service nova] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Received unexpected event network-vif-plugged-6fcb037d-fd48-490d-9898-b952ffaad4ef for instance with vm_state building and task_state spawning. [ 822.932159] env[61987]: DEBUG nova.compute.manager [req-e8671de3-9c14-4b55-b238-468574336d59 req-67efda73-049d-4127-86fb-a80b778e8dbd service nova] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Received event network-changed-6fcb037d-fd48-490d-9898-b952ffaad4ef {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 822.932159] env[61987]: DEBUG nova.compute.manager [req-e8671de3-9c14-4b55-b238-468574336d59 req-67efda73-049d-4127-86fb-a80b778e8dbd service nova] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Refreshing instance network info cache due to event network-changed-6fcb037d-fd48-490d-9898-b952ffaad4ef. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 822.932159] env[61987]: DEBUG oslo_concurrency.lockutils [req-e8671de3-9c14-4b55-b238-468574336d59 req-67efda73-049d-4127-86fb-a80b778e8dbd service nova] Acquiring lock "refresh_cache-93706a54-1fc4-4aed-8807-ec2c3d3694b4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.979972] env[61987]: DEBUG oslo_concurrency.lockutils [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Releasing lock "refresh_cache-93706a54-1fc4-4aed-8807-ec2c3d3694b4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.979972] env[61987]: DEBUG nova.compute.manager [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Instance network_info: |[{"id": "6fcb037d-fd48-490d-9898-b952ffaad4ef", "address": "fa:16:3e:14:a7:3a", "network": {"id": "35c53c0f-af9d-4934-84c7-ea4c5252fa58", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-343092073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d32dc833164a428dad56d1eeb9a7c3eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "790c811b-3fa6-49f8-87ac-c51450911137", "external-id": "nsx-vlan-transportzone-908", "segmentation_id": 908, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fcb037d-fd", "ovs_interfaceid": "6fcb037d-fd48-490d-9898-b952ffaad4ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 822.979972] env[61987]: DEBUG oslo_concurrency.lockutils [req-e8671de3-9c14-4b55-b238-468574336d59 req-67efda73-049d-4127-86fb-a80b778e8dbd service nova] Acquired lock "refresh_cache-93706a54-1fc4-4aed-8807-ec2c3d3694b4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.979972] env[61987]: DEBUG nova.network.neutron [req-e8671de3-9c14-4b55-b238-468574336d59 req-67efda73-049d-4127-86fb-a80b778e8dbd service nova] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Refreshing network info cache for port 6fcb037d-fd48-490d-9898-b952ffaad4ef {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 822.980346] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:a7:3a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '790c811b-3fa6-49f8-87ac-c51450911137', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6fcb037d-fd48-490d-9898-b952ffaad4ef', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 822.994202] env[61987]: DEBUG oslo.service.loopingcall [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.994202] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 822.994202] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6a78609a-b41b-4517-a74a-884a7fc9b969 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.016274] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 823.016274] env[61987]: value = "task-1061700" [ 823.016274] env[61987]: _type = "Task" [ 823.016274] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.026400] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061700, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.145023] env[61987]: INFO nova.compute.manager [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Rebuilding instance [ 823.186857] env[61987]: DEBUG nova.scheduler.client.report [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 823.196700] env[61987]: DEBUG nova.compute.manager [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 823.197938] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb99538-c97b-428e-9655-df2c23bb07fe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.204551] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': task-1061699, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481816} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.206239] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] fe22fc1b-61fb-4088-b5fe-b84eb4ec4099/fe22fc1b-61fb-4088-b5fe-b84eb4ec4099.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 823.206874] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 823.207603] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-65582867-84c8-4e5e-90c9-68acd0ef224f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.222130] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Waiting for the task: (returnval){ [ 823.222130] env[61987]: value = "task-1061701" [ 823.222130] env[61987]: _type = "Task" [ 823.222130] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.231300] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': task-1061701, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.242025] env[61987]: DEBUG nova.compute.manager [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 823.266104] env[61987]: DEBUG nova.virt.hardware [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 823.266104] env[61987]: DEBUG nova.virt.hardware [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 823.266104] env[61987]: DEBUG nova.virt.hardware [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 823.266104] env[61987]: DEBUG nova.virt.hardware [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 823.266104] env[61987]: DEBUG nova.virt.hardware [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 823.266104] env[61987]: DEBUG nova.virt.hardware [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 823.266398] env[61987]: DEBUG nova.virt.hardware [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 823.266398] env[61987]: DEBUG nova.virt.hardware [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 823.266530] env[61987]: DEBUG nova.virt.hardware [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 823.266739] env[61987]: DEBUG nova.virt.hardware [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 823.268014] env[61987]: DEBUG nova.virt.hardware [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 823.268014] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e83ef772-3937-4aba-b4a1-5b2321c073dd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.275871] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2afea02-c70f-4f7c-87ef-0f746d7c5095 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.529908] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061700, 'name': CreateVM_Task, 'duration_secs': 0.442277} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.530290] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 823.531551] env[61987]: DEBUG oslo_concurrency.lockutils [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.531855] env[61987]: DEBUG oslo_concurrency.lockutils [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.532288] env[61987]: DEBUG oslo_concurrency.lockutils [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 823.532658] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dabfd890-1a26-4678-9bda-2bdf6ff3f73a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.538200] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 823.538200] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52fc17a7-88a9-f68c-0fe0-464ddcea267d" [ 823.538200] env[61987]: _type = "Task" [ 823.538200] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.547863] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52fc17a7-88a9-f68c-0fe0-464ddcea267d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.693052] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.477s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.694070] env[61987]: DEBUG nova.compute.manager [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 823.696255] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.416s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.696504] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.698329] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.157s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.699715] env[61987]: INFO nova.compute.claims [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.730458] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': task-1061701, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.09549} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.730739] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 823.731678] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fc796ab-fed2-459f-a87f-1d7c34be77ae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.734975] env[61987]: INFO nova.scheduler.client.report [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Deleted allocations for instance 93a27dea-f28a-4cbf-ac37-ff41763ae18d [ 823.756270] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] fe22fc1b-61fb-4088-b5fe-b84eb4ec4099/fe22fc1b-61fb-4088-b5fe-b84eb4ec4099.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 823.757771] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d289a465-8e6e-4e87-90cc-5537ab8895aa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.780657] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Waiting for the task: (returnval){ [ 823.780657] env[61987]: value = "task-1061702" [ 823.780657] env[61987]: _type = "Task" [ 823.780657] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.790587] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': task-1061702, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.948227] env[61987]: DEBUG nova.network.neutron [req-e8671de3-9c14-4b55-b238-468574336d59 req-67efda73-049d-4127-86fb-a80b778e8dbd service nova] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Updated VIF entry in instance network info cache for port 6fcb037d-fd48-490d-9898-b952ffaad4ef. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 823.948614] env[61987]: DEBUG nova.network.neutron [req-e8671de3-9c14-4b55-b238-468574336d59 req-67efda73-049d-4127-86fb-a80b778e8dbd service nova] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Updating instance_info_cache with network_info: [{"id": "6fcb037d-fd48-490d-9898-b952ffaad4ef", "address": "fa:16:3e:14:a7:3a", "network": {"id": "35c53c0f-af9d-4934-84c7-ea4c5252fa58", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-343092073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d32dc833164a428dad56d1eeb9a7c3eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "790c811b-3fa6-49f8-87ac-c51450911137", "external-id": "nsx-vlan-transportzone-908", "segmentation_id": 908, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fcb037d-fd", "ovs_interfaceid": "6fcb037d-fd48-490d-9898-b952ffaad4ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.050842] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52fc17a7-88a9-f68c-0fe0-464ddcea267d, 'name': SearchDatastore_Task, 'duration_secs': 0.009699} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.053239] env[61987]: DEBUG oslo_concurrency.lockutils [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.054227] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 824.056701] env[61987]: DEBUG oslo_concurrency.lockutils [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.056701] env[61987]: DEBUG oslo_concurrency.lockutils [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.056701] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 824.056701] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8296ea8d-4771-47e0-a649-2b4355244373 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.066024] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 824.066024] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 824.067613] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6dbca63c-2658-4d81-88cc-23cbcc038263 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.077969] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 824.077969] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52c0230e-d072-2011-0874-6365ed876711" [ 824.077969] env[61987]: _type = "Task" [ 824.077969] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.090801] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c0230e-d072-2011-0874-6365ed876711, 'name': SearchDatastore_Task, 'duration_secs': 0.009534} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.091745] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d20c69d-3c76-49f8-a892-f247e70fd4ca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.098017] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 824.098017] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5221f638-970e-668e-e1bc-a6733b13ad86" [ 824.098017] env[61987]: _type = "Task" [ 824.098017] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.105407] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5221f638-970e-668e-e1bc-a6733b13ad86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.204154] env[61987]: DEBUG nova.compute.utils [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 824.205714] env[61987]: DEBUG nova.compute.manager [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 824.205882] env[61987]: DEBUG nova.network.neutron [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 824.219565] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 824.219850] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-404205b2-1b36-48a1-91a1-a13d72021705 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.229023] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for the task: (returnval){ [ 824.229023] env[61987]: value = "task-1061703" [ 824.229023] env[61987]: _type = "Task" [ 824.229023] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.239220] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061703, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.257294] env[61987]: DEBUG nova.network.neutron [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Successfully updated port: 69ce8bb3-dc28-422e-96a8-833062bd6b77 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 824.261213] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ca99cc9b-481a-4157-aba3-c3e7de657dc6 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "93a27dea-f28a-4cbf-ac37-ff41763ae18d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.611s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.275521] env[61987]: DEBUG nova.policy [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0f84a1e7ec448fe9d14f7b909e1952a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '48d633a419cd468dae6134421c9c17fa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 824.290135] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': task-1061702, 'name': ReconfigVM_Task, 'duration_secs': 0.303381} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.290620] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Reconfigured VM instance instance-00000041 to attach disk [datastore1] fe22fc1b-61fb-4088-b5fe-b84eb4ec4099/fe22fc1b-61fb-4088-b5fe-b84eb4ec4099.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 824.291166] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-969fe1c4-d619-4b1e-84ab-e09002659c6d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.297466] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Waiting for the task: (returnval){ [ 824.297466] env[61987]: value = "task-1061704" [ 824.297466] env[61987]: _type = "Task" [ 824.297466] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.306240] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': task-1061704, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.451906] env[61987]: DEBUG oslo_concurrency.lockutils [req-e8671de3-9c14-4b55-b238-468574336d59 req-67efda73-049d-4127-86fb-a80b778e8dbd service nova] Releasing lock "refresh_cache-93706a54-1fc4-4aed-8807-ec2c3d3694b4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.572340] env[61987]: DEBUG nova.network.neutron [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Successfully created port: 6119ef11-a392-4c05-8c71-6e5b25f3cdc9 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 824.615642] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5221f638-970e-668e-e1bc-a6733b13ad86, 'name': SearchDatastore_Task, 'duration_secs': 0.007976} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.616599] env[61987]: DEBUG oslo_concurrency.lockutils [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.616599] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 93706a54-1fc4-4aed-8807-ec2c3d3694b4/93706a54-1fc4-4aed-8807-ec2c3d3694b4.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 824.616599] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b7a0c0b6-89ab-4d2b-8c5b-916f4dff35c5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.623182] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 824.623182] env[61987]: value = "task-1061705" [ 824.623182] env[61987]: _type = "Task" [ 824.623182] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.631113] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061705, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.712074] env[61987]: DEBUG nova.compute.manager [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 824.738235] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061703, 'name': PowerOffVM_Task, 'duration_secs': 0.118397} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.738554] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 824.738789] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 824.741981] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e226553-0cbf-40e7-bcf9-071b67ba7b5d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.749137] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 824.749422] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4e26af09-511e-4beb-a763-27ea00256178 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.761699] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Acquiring lock "refresh_cache-98d895ee-1d1f-49cd-a055-fe6c8462a692" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.761832] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Acquired lock "refresh_cache-98d895ee-1d1f-49cd-a055-fe6c8462a692" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.761890] env[61987]: DEBUG nova.network.neutron [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 824.778704] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 824.781020] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 824.781020] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Deleting the datastore file [datastore2] 59d34772-26a5-4af2-82dc-da932e996745 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 824.781020] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1a98c0f6-f1d2-451e-8a81-24320d5be2da {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.787733] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for the task: (returnval){ [ 824.787733] env[61987]: value = "task-1061707" [ 824.787733] env[61987]: _type = "Task" [ 824.787733] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.799252] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061707, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.807667] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': task-1061704, 'name': Rename_Task, 'duration_secs': 0.147256} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.810467] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 824.811086] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-70ffa195-3e94-4d43-9e80-10881347a122 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.817876] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Waiting for the task: (returnval){ [ 824.817876] env[61987]: value = "task-1061708" [ 824.817876] env[61987]: _type = "Task" [ 824.817876] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.827797] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': task-1061708, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.069431] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6046b7e5-b3eb-4ad6-95cc-351655e74ad7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.078167] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8adaebd-c25c-4355-917e-a05c811e3354 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.114408] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d33f1d8e-72b1-437d-99a7-28394eb5755a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.118603] env[61987]: DEBUG nova.compute.manager [req-f8a6b1b0-f704-4199-a438-b5a7dd00183e req-080df506-98e5-4c19-ac10-74b50974f433 service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Received event network-vif-plugged-69ce8bb3-dc28-422e-96a8-833062bd6b77 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 825.118841] env[61987]: DEBUG oslo_concurrency.lockutils [req-f8a6b1b0-f704-4199-a438-b5a7dd00183e req-080df506-98e5-4c19-ac10-74b50974f433 service nova] Acquiring lock "98d895ee-1d1f-49cd-a055-fe6c8462a692-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.119101] env[61987]: DEBUG oslo_concurrency.lockutils [req-f8a6b1b0-f704-4199-a438-b5a7dd00183e req-080df506-98e5-4c19-ac10-74b50974f433 service nova] Lock "98d895ee-1d1f-49cd-a055-fe6c8462a692-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.119253] env[61987]: DEBUG oslo_concurrency.lockutils [req-f8a6b1b0-f704-4199-a438-b5a7dd00183e req-080df506-98e5-4c19-ac10-74b50974f433 service nova] Lock "98d895ee-1d1f-49cd-a055-fe6c8462a692-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.119623] env[61987]: DEBUG nova.compute.manager [req-f8a6b1b0-f704-4199-a438-b5a7dd00183e req-080df506-98e5-4c19-ac10-74b50974f433 service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] No waiting events found dispatching network-vif-plugged-69ce8bb3-dc28-422e-96a8-833062bd6b77 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 825.119623] env[61987]: WARNING nova.compute.manager [req-f8a6b1b0-f704-4199-a438-b5a7dd00183e req-080df506-98e5-4c19-ac10-74b50974f433 service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Received unexpected event network-vif-plugged-69ce8bb3-dc28-422e-96a8-833062bd6b77 for instance with vm_state building and task_state spawning. [ 825.119767] env[61987]: DEBUG nova.compute.manager [req-f8a6b1b0-f704-4199-a438-b5a7dd00183e req-080df506-98e5-4c19-ac10-74b50974f433 service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Received event network-changed-69ce8bb3-dc28-422e-96a8-833062bd6b77 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 825.120555] env[61987]: DEBUG nova.compute.manager [req-f8a6b1b0-f704-4199-a438-b5a7dd00183e req-080df506-98e5-4c19-ac10-74b50974f433 service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Refreshing instance network info cache due to event network-changed-69ce8bb3-dc28-422e-96a8-833062bd6b77. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 825.120555] env[61987]: DEBUG oslo_concurrency.lockutils [req-f8a6b1b0-f704-4199-a438-b5a7dd00183e req-080df506-98e5-4c19-ac10-74b50974f433 service nova] Acquiring lock "refresh_cache-98d895ee-1d1f-49cd-a055-fe6c8462a692" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.129674] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce6ae46-e82f-4b3c-9570-d933cb3c34ff {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.140307] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061705, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.149519] env[61987]: DEBUG nova.compute.provider_tree [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 825.301731] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061707, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.121703} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.302549] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 825.302549] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 825.302549] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 825.314035] env[61987]: DEBUG nova.network.neutron [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 825.337253] env[61987]: DEBUG oslo_vmware.api [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': task-1061708, 'name': PowerOnVM_Task, 'duration_secs': 0.510553} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.338151] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 825.338151] env[61987]: INFO nova.compute.manager [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Took 6.98 seconds to spawn the instance on the hypervisor. [ 825.338427] env[61987]: DEBUG nova.compute.manager [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 825.341171] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a0f8a4-30f8-4840-afd8-22a4da7f3b30 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.477526] env[61987]: DEBUG nova.network.neutron [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Updating instance_info_cache with network_info: [{"id": "69ce8bb3-dc28-422e-96a8-833062bd6b77", "address": "fa:16:3e:ab:10:c2", "network": {"id": "35162a8d-60d3-465b-9330-64bbd235d6a2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-981766067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8dffc4d03e514fb8932b80718375084b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69ce8bb3-dc", "ovs_interfaceid": "69ce8bb3-dc28-422e-96a8-833062bd6b77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.635632] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061705, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.553176} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.636095] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 93706a54-1fc4-4aed-8807-ec2c3d3694b4/93706a54-1fc4-4aed-8807-ec2c3d3694b4.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 825.636488] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 825.637113] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bebd29dd-82f5-4a62-909d-25e622cfa378 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.645546] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 825.645546] env[61987]: value = "task-1061709" [ 825.645546] env[61987]: _type = "Task" [ 825.645546] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.654229] env[61987]: DEBUG nova.scheduler.client.report [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 825.657442] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061709, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.725311] env[61987]: DEBUG nova.compute.manager [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 825.751816] env[61987]: DEBUG nova.virt.hardware [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 825.752123] env[61987]: DEBUG nova.virt.hardware [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 825.752297] env[61987]: DEBUG nova.virt.hardware [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 825.752489] env[61987]: DEBUG nova.virt.hardware [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 825.752642] env[61987]: DEBUG nova.virt.hardware [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 825.752794] env[61987]: DEBUG nova.virt.hardware [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 825.752998] env[61987]: DEBUG nova.virt.hardware [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 825.753180] env[61987]: DEBUG nova.virt.hardware [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 825.753354] env[61987]: DEBUG nova.virt.hardware [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 825.753519] env[61987]: DEBUG nova.virt.hardware [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 825.753697] env[61987]: DEBUG nova.virt.hardware [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 825.754609] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bef92db-7c94-4d2f-b21b-b13006ce1674 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.762662] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4c731cc-d5be-4a0c-8e12-55d2793faeb2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.858618] env[61987]: INFO nova.compute.manager [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Took 46.51 seconds to build instance. [ 825.980143] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Releasing lock "refresh_cache-98d895ee-1d1f-49cd-a055-fe6c8462a692" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.980546] env[61987]: DEBUG nova.compute.manager [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Instance network_info: |[{"id": "69ce8bb3-dc28-422e-96a8-833062bd6b77", "address": "fa:16:3e:ab:10:c2", "network": {"id": "35162a8d-60d3-465b-9330-64bbd235d6a2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-981766067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8dffc4d03e514fb8932b80718375084b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69ce8bb3-dc", "ovs_interfaceid": "69ce8bb3-dc28-422e-96a8-833062bd6b77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 825.980877] env[61987]: DEBUG oslo_concurrency.lockutils [req-f8a6b1b0-f704-4199-a438-b5a7dd00183e req-080df506-98e5-4c19-ac10-74b50974f433 service nova] Acquired lock "refresh_cache-98d895ee-1d1f-49cd-a055-fe6c8462a692" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.981107] env[61987]: DEBUG nova.network.neutron [req-f8a6b1b0-f704-4199-a438-b5a7dd00183e req-080df506-98e5-4c19-ac10-74b50974f433 service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Refreshing network info cache for port 69ce8bb3-dc28-422e-96a8-833062bd6b77 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 825.982313] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ab:10:c2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '68add7d6-c025-46fa-84d3-9c589adb63e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '69ce8bb3-dc28-422e-96a8-833062bd6b77', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 825.989866] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Creating folder: Project (8dffc4d03e514fb8932b80718375084b). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 825.990352] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-91f4d628-3c1b-4a01-b690-b9cf58940925 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.001993] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Created folder: Project (8dffc4d03e514fb8932b80718375084b) in parent group-v234219. [ 826.002241] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Creating folder: Instances. Parent ref: group-v234345. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 826.002802] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d50dd5d8-e69e-467d-8869-0a980b800d30 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.011474] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Created folder: Instances in parent group-v234345. [ 826.011592] env[61987]: DEBUG oslo.service.loopingcall [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 826.011777] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 826.011978] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-54e4f90e-1aa1-4c65-ad09-c90822401650 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.029955] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 826.029955] env[61987]: value = "task-1061712" [ 826.029955] env[61987]: _type = "Task" [ 826.029955] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.036878] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061712, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.140870] env[61987]: DEBUG nova.compute.manager [req-a1a7b81f-7595-4905-bf9a-f65eb641af9d req-aa74e759-ff31-4a16-ae53-49063aa6eec6 service nova] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Received event network-vif-plugged-6119ef11-a392-4c05-8c71-6e5b25f3cdc9 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 826.141129] env[61987]: DEBUG oslo_concurrency.lockutils [req-a1a7b81f-7595-4905-bf9a-f65eb641af9d req-aa74e759-ff31-4a16-ae53-49063aa6eec6 service nova] Acquiring lock "6ed3f0e9-3c39-4406-9863-931a117961a8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.141353] env[61987]: DEBUG oslo_concurrency.lockutils [req-a1a7b81f-7595-4905-bf9a-f65eb641af9d req-aa74e759-ff31-4a16-ae53-49063aa6eec6 service nova] Lock "6ed3f0e9-3c39-4406-9863-931a117961a8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.141740] env[61987]: DEBUG oslo_concurrency.lockutils [req-a1a7b81f-7595-4905-bf9a-f65eb641af9d req-aa74e759-ff31-4a16-ae53-49063aa6eec6 service nova] Lock "6ed3f0e9-3c39-4406-9863-931a117961a8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.141954] env[61987]: DEBUG nova.compute.manager [req-a1a7b81f-7595-4905-bf9a-f65eb641af9d req-aa74e759-ff31-4a16-ae53-49063aa6eec6 service nova] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] No waiting events found dispatching network-vif-plugged-6119ef11-a392-4c05-8c71-6e5b25f3cdc9 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 826.142194] env[61987]: WARNING nova.compute.manager [req-a1a7b81f-7595-4905-bf9a-f65eb641af9d req-aa74e759-ff31-4a16-ae53-49063aa6eec6 service nova] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Received unexpected event network-vif-plugged-6119ef11-a392-4c05-8c71-6e5b25f3cdc9 for instance with vm_state building and task_state spawning. [ 826.154734] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061709, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067275} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.154981] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 826.156059] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-febabb44-9ad4-448d-b6e8-4e35b5ed830b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.161244] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.463s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.161929] env[61987]: DEBUG nova.compute.manager [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 826.164497] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.974s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.164689] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.164864] env[61987]: INFO nova.compute.manager [None req-6f23ffe0-be68-45bd-9bda-41ad7197e449 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Successfully reverted task state from rebuilding on failure for instance. [ 826.170588] env[61987]: DEBUG oslo_concurrency.lockutils [None req-acb92527-6504-4fe5-bca2-a703c7bd5326 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.780s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.170832] env[61987]: DEBUG oslo_concurrency.lockutils [None req-acb92527-6504-4fe5-bca2-a703c7bd5326 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.172953] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.239s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.173159] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.175424] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.331s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.177206] env[61987]: INFO nova.compute.claims [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 826.217953] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 93706a54-1fc4-4aed-8807-ec2c3d3694b4/93706a54-1fc4-4aed-8807-ec2c3d3694b4.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 826.222461] env[61987]: INFO nova.scheduler.client.report [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Deleted allocations for instance 10354ba0-5a64-4174-9c61-152d54a5dfcc [ 826.223990] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dab7ede3-ec40-4df9-96ef-2dde06aa4b58 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.254635] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 826.254635] env[61987]: value = "task-1061713" [ 826.254635] env[61987]: _type = "Task" [ 826.254635] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.266031] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061713, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.289779] env[61987]: DEBUG nova.network.neutron [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Successfully updated port: 6119ef11-a392-4c05-8c71-6e5b25f3cdc9 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 826.339877] env[61987]: DEBUG nova.virt.hardware [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 826.340190] env[61987]: DEBUG nova.virt.hardware [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 826.340371] env[61987]: DEBUG nova.virt.hardware [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 826.340582] env[61987]: DEBUG nova.virt.hardware [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 826.340766] env[61987]: DEBUG nova.virt.hardware [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 826.340940] env[61987]: DEBUG nova.virt.hardware [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 826.341320] env[61987]: DEBUG nova.virt.hardware [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 826.341530] env[61987]: DEBUG nova.virt.hardware [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 826.341732] env[61987]: DEBUG nova.virt.hardware [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 826.341944] env[61987]: DEBUG nova.virt.hardware [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 826.342163] env[61987]: DEBUG nova.virt.hardware [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 826.343148] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1cfff97-ad7e-4451-b037-aff8cff95f51 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.352054] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a26bba79-f635-4225-9037-19ff56cd7425 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.366802] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5bd9d59-802c-4e5b-aa36-0c6a99ac31c4 tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Lock "fe22fc1b-61fb-4088-b5fe-b84eb4ec4099" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.029s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.367282] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Instance VIF info [] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 826.373031] env[61987]: DEBUG oslo.service.loopingcall [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 826.373341] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 826.373586] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f49c1ae0-f39d-45be-8e03-095fac63d663 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.390683] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 826.390683] env[61987]: value = "task-1061714" [ 826.390683] env[61987]: _type = "Task" [ 826.390683] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.400032] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061714, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.541226] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061712, 'name': CreateVM_Task, 'duration_secs': 0.441677} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.541226] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 826.541784] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.541988] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.542497] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 826.542776] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32eeca61-bcd4-4d94-82bf-b49f925f99ae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.547810] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Waiting for the task: (returnval){ [ 826.547810] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5244493b-eef5-b964-635e-0dfd20f64660" [ 826.547810] env[61987]: _type = "Task" [ 826.547810] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.556268] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5244493b-eef5-b964-635e-0dfd20f64660, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.682976] env[61987]: DEBUG nova.compute.utils [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 826.683945] env[61987]: DEBUG nova.compute.manager [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 826.684374] env[61987]: DEBUG nova.network.neutron [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 826.757344] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a016e3e9-d125-4faa-abca-af8266fabd84 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "10354ba0-5a64-4174-9c61-152d54a5dfcc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.797s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.761803] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ec62fa3-5350-4b50-97dc-e34715d9544d tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "10354ba0-5a64-4174-9c61-152d54a5dfcc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 25.480s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.761803] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ec62fa3-5350-4b50-97dc-e34715d9544d tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquiring lock "10354ba0-5a64-4174-9c61-152d54a5dfcc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.761803] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ec62fa3-5350-4b50-97dc-e34715d9544d tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "10354ba0-5a64-4174-9c61-152d54a5dfcc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.761803] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ec62fa3-5350-4b50-97dc-e34715d9544d tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "10354ba0-5a64-4174-9c61-152d54a5dfcc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.762176] env[61987]: DEBUG nova.policy [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c124c6b758543a68e141796b585e41f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bd47eb44bc334bf3ae5813905903ecbc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 826.767549] env[61987]: INFO nova.compute.manager [None req-8ec62fa3-5350-4b50-97dc-e34715d9544d tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Terminating instance [ 826.776101] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061713, 'name': ReconfigVM_Task, 'duration_secs': 0.270037} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.776756] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 93706a54-1fc4-4aed-8807-ec2c3d3694b4/93706a54-1fc4-4aed-8807-ec2c3d3694b4.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 826.777186] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Creating Virtual Disk of size 1048576 KB and adapter type paraVirtual on the data store {{(pid=61987) create_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1373}} [ 826.778072] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CreateVirtualDisk_Task with opID=oslo.vmware-c47b666f-8b36-42bf-b2bd-fed495686b49 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.782612] env[61987]: DEBUG oslo_concurrency.lockutils [None req-acb92527-6504-4fe5-bca2-a703c7bd5326 tempest-ServerActionsV293TestJSON-1814806386 tempest-ServerActionsV293TestJSON-1814806386-project-member] Lock "95641d0b-970c-4b94-8568-5c46d0808345" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.233s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.786761] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 826.786761] env[61987]: value = "task-1061715" [ 826.786761] env[61987]: _type = "Task" [ 826.786761] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.794031] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061715, 'name': CreateVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.797976] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Acquiring lock "refresh_cache-6ed3f0e9-3c39-4406-9863-931a117961a8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.798229] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Acquired lock "refresh_cache-6ed3f0e9-3c39-4406-9863-931a117961a8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.798342] env[61987]: DEBUG nova.network.neutron [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 826.900612] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061714, 'name': CreateVM_Task, 'duration_secs': 0.286379} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.900783] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 826.901212] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.059159] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5244493b-eef5-b964-635e-0dfd20f64660, 'name': SearchDatastore_Task, 'duration_secs': 0.009392} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.059490] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.060327] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 827.060327] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.060327] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.060327] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 827.060579] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.060900] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 827.061181] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e635dd04-4a79-47dd-8ffb-bfb26170d0df {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.063117] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b444e5e-47e8-49b0-b73c-5cee3b72184a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.068448] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for the task: (returnval){ [ 827.068448] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]523f845e-61cc-f4df-5dbf-b91e7c0e799f" [ 827.068448] env[61987]: _type = "Task" [ 827.068448] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.072450] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 827.072631] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 827.073680] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-841d2c5d-19c7-4a42-8d98-09df764fe3a9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.078914] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523f845e-61cc-f4df-5dbf-b91e7c0e799f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.081878] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Waiting for the task: (returnval){ [ 827.081878] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52e50ee4-0209-9442-12c5-e26ceb80f666" [ 827.081878] env[61987]: _type = "Task" [ 827.081878] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.090777] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52e50ee4-0209-9442-12c5-e26ceb80f666, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.161494] env[61987]: DEBUG nova.network.neutron [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Successfully created port: e88ac5c4-67fa-4013-ac31-ca59251955b5 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 827.187756] env[61987]: DEBUG nova.compute.manager [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 827.203168] env[61987]: DEBUG nova.network.neutron [req-f8a6b1b0-f704-4199-a438-b5a7dd00183e req-080df506-98e5-4c19-ac10-74b50974f433 service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Updated VIF entry in instance network info cache for port 69ce8bb3-dc28-422e-96a8-833062bd6b77. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 827.203257] env[61987]: DEBUG nova.network.neutron [req-f8a6b1b0-f704-4199-a438-b5a7dd00183e req-080df506-98e5-4c19-ac10-74b50974f433 service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Updating instance_info_cache with network_info: [{"id": "69ce8bb3-dc28-422e-96a8-833062bd6b77", "address": "fa:16:3e:ab:10:c2", "network": {"id": "35162a8d-60d3-465b-9330-64bbd235d6a2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-981766067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8dffc4d03e514fb8932b80718375084b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69ce8bb3-dc", "ovs_interfaceid": "69ce8bb3-dc28-422e-96a8-833062bd6b77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.271624] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ec62fa3-5350-4b50-97dc-e34715d9544d tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquiring lock "refresh_cache-10354ba0-5a64-4174-9c61-152d54a5dfcc" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.271823] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ec62fa3-5350-4b50-97dc-e34715d9544d tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquired lock "refresh_cache-10354ba0-5a64-4174-9c61-152d54a5dfcc" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.272021] env[61987]: DEBUG nova.network.neutron [None req-8ec62fa3-5350-4b50-97dc-e34715d9544d tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 827.300040] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061715, 'name': CreateVirtualDisk_Task, 'duration_secs': 0.053175} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.302026] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Created Virtual Disk of size 1048576 KB and type thin {{(pid=61987) create_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1394}} [ 827.303148] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-549753e1-4f56-47eb-9cd0-3a5d95733489 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.335322] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 93706a54-1fc4-4aed-8807-ec2c3d3694b4/ephemeral_0.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 827.338341] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8361526-bbfe-44fe-994c-ad227f20c5de {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.362891] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 827.362891] env[61987]: value = "task-1061716" [ 827.362891] env[61987]: _type = "Task" [ 827.362891] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.374626] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061716, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.383281] env[61987]: DEBUG nova.network.neutron [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 827.583837] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523f845e-61cc-f4df-5dbf-b91e7c0e799f, 'name': SearchDatastore_Task, 'duration_secs': 0.009169} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.586760] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.587593] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 827.587593] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.594390] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52e50ee4-0209-9442-12c5-e26ceb80f666, 'name': SearchDatastore_Task, 'duration_secs': 0.008537} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.596539] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13df882a-94ad-4be2-b7cb-a2bccf2734a3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.599558] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10e13117-15cd-40dd-9138-137b480b4dac {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.607401] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a186da9c-9c65-4e3d-9977-b73a409f2213 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.610650] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Waiting for the task: (returnval){ [ 827.610650] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]520dadeb-8972-8a32-3e95-2118fda0d5f1" [ 827.610650] env[61987]: _type = "Task" [ 827.610650] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.642191] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a64e5a86-fb8e-4116-bd71-e61407196f5a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.648659] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]520dadeb-8972-8a32-3e95-2118fda0d5f1, 'name': SearchDatastore_Task, 'duration_secs': 0.010409} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.649321] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.649610] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 98d895ee-1d1f-49cd-a055-fe6c8462a692/98d895ee-1d1f-49cd-a055-fe6c8462a692.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 827.649924] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.650107] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 827.650329] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f52365dd-7d27-4de9-bfcc-6ec0df09459b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.654195] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-09047306-37b1-46af-b1ed-6e19db45fbaf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.657194] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51cc34a2-7b36-408e-9c0c-7c826a49c41d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.664358] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Waiting for the task: (returnval){ [ 827.664358] env[61987]: value = "task-1061717" [ 827.664358] env[61987]: _type = "Task" [ 827.664358] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.681827] env[61987]: DEBUG nova.compute.provider_tree [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 827.681827] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 827.681980] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 827.683302] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2c9c5ce-30c5-42d0-bc56-355d0021f93c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.688986] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061717, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.696022] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for the task: (returnval){ [ 827.696022] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5200298d-3c60-36a6-a817-faf8b68d9249" [ 827.696022] env[61987]: _type = "Task" [ 827.696022] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.704908] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5200298d-3c60-36a6-a817-faf8b68d9249, 'name': SearchDatastore_Task, 'duration_secs': 0.009257} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.705391] env[61987]: DEBUG oslo_concurrency.lockutils [req-f8a6b1b0-f704-4199-a438-b5a7dd00183e req-080df506-98e5-4c19-ac10-74b50974f433 service nova] Releasing lock "refresh_cache-98d895ee-1d1f-49cd-a055-fe6c8462a692" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.706366] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb721b73-7199-4b6c-853e-8c31f078a056 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.714423] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for the task: (returnval){ [ 827.714423] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5282e539-6dbc-59a3-f602-ace3ec4fd31b" [ 827.714423] env[61987]: _type = "Task" [ 827.714423] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.724550] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5282e539-6dbc-59a3-f602-ace3ec4fd31b, 'name': SearchDatastore_Task, 'duration_secs': 0.008915} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.725386] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.725629] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 59d34772-26a5-4af2-82dc-da932e996745/59d34772-26a5-4af2-82dc-da932e996745.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 827.725979] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-603d2a96-5dbc-45a8-ae97-f078254a3891 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.733147] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for the task: (returnval){ [ 827.733147] env[61987]: value = "task-1061718" [ 827.733147] env[61987]: _type = "Task" [ 827.733147] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.741740] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061718, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.742510] env[61987]: DEBUG nova.network.neutron [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Updating instance_info_cache with network_info: [{"id": "6119ef11-a392-4c05-8c71-6e5b25f3cdc9", "address": "fa:16:3e:cc:5c:78", "network": {"id": "a91d7bba-6d0b-4ee1-a490-ea769b3b83ca", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-546229287-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48d633a419cd468dae6134421c9c17fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "72781990-3cb3-42eb-9eb1-4040dedbf66f", "external-id": "cl2-zone-812", "segmentation_id": 812, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6119ef11-a3", "ovs_interfaceid": "6119ef11-a392-4c05-8c71-6e5b25f3cdc9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.774711] env[61987]: DEBUG nova.compute.utils [None req-8ec62fa3-5350-4b50-97dc-e34715d9544d tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Can not refresh info_cache because instance was not found {{(pid=61987) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1055}} [ 827.798069] env[61987]: DEBUG nova.network.neutron [None req-8ec62fa3-5350-4b50-97dc-e34715d9544d tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 827.877717] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061716, 'name': ReconfigVM_Task, 'duration_secs': 0.495318} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.878041] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 93706a54-1fc4-4aed-8807-ec2c3d3694b4/ephemeral_0.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 827.878742] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1e0be28d-413e-41ad-9f59-e3ecc6c69285 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.886685] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 827.886685] env[61987]: value = "task-1061719" [ 827.886685] env[61987]: _type = "Task" [ 827.886685] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.895905] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061719, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.930744] env[61987]: DEBUG nova.network.neutron [None req-8ec62fa3-5350-4b50-97dc-e34715d9544d tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.179116] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061717, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473699} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.179240] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 98d895ee-1d1f-49cd-a055-fe6c8462a692/98d895ee-1d1f-49cd-a055-fe6c8462a692.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 828.179471] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.179739] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b7a9ef4d-0fbe-468a-8011-a8370359d4f4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.183982] env[61987]: DEBUG nova.scheduler.client.report [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 828.191022] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Waiting for the task: (returnval){ [ 828.191022] env[61987]: value = "task-1061720" [ 828.191022] env[61987]: _type = "Task" [ 828.191022] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.200780] env[61987]: DEBUG nova.compute.manager [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 828.203541] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061720, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.215457] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquiring lock "4edd0159-6db1-41bd-a616-31a34e984059" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.215754] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "4edd0159-6db1-41bd-a616-31a34e984059" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.215976] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquiring lock "4edd0159-6db1-41bd-a616-31a34e984059-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.216639] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "4edd0159-6db1-41bd-a616-31a34e984059-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.216850] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "4edd0159-6db1-41bd-a616-31a34e984059-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.219364] env[61987]: INFO nova.compute.manager [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Terminating instance [ 828.232971] env[61987]: DEBUG nova.virt.hardware [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 828.233257] env[61987]: DEBUG nova.virt.hardware [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.233426] env[61987]: DEBUG nova.virt.hardware [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 828.233621] env[61987]: DEBUG nova.virt.hardware [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.233782] env[61987]: DEBUG nova.virt.hardware [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 828.234011] env[61987]: DEBUG nova.virt.hardware [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 828.234511] env[61987]: DEBUG nova.virt.hardware [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 828.234511] env[61987]: DEBUG nova.virt.hardware [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 828.234750] env[61987]: DEBUG nova.virt.hardware [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 828.234945] env[61987]: DEBUG nova.virt.hardware [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 828.235152] env[61987]: DEBUG nova.virt.hardware [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 828.236511] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed6c170a-bc62-4906-b432-89de6827f1e5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.244906] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Releasing lock "refresh_cache-6ed3f0e9-3c39-4406-9863-931a117961a8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.245252] env[61987]: DEBUG nova.compute.manager [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Instance network_info: |[{"id": "6119ef11-a392-4c05-8c71-6e5b25f3cdc9", "address": "fa:16:3e:cc:5c:78", "network": {"id": "a91d7bba-6d0b-4ee1-a490-ea769b3b83ca", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-546229287-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48d633a419cd468dae6134421c9c17fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "72781990-3cb3-42eb-9eb1-4040dedbf66f", "external-id": "cl2-zone-812", "segmentation_id": 812, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6119ef11-a3", "ovs_interfaceid": "6119ef11-a392-4c05-8c71-6e5b25f3cdc9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 828.245667] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cc:5c:78', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '72781990-3cb3-42eb-9eb1-4040dedbf66f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6119ef11-a392-4c05-8c71-6e5b25f3cdc9', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 828.253422] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Creating folder: Project (48d633a419cd468dae6134421c9c17fa). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 828.259781] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5226f7d9-ad5c-4c8a-8dcb-878b35eaa703 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.262215] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061718, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.264854] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf467bce-1015-47ae-9100-815d30b84061 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.273311] env[61987]: DEBUG nova.compute.manager [req-db02cab5-8880-4fea-b11c-8feeb401b63c req-81d71756-7152-4601-b4db-9a4d6504332b service nova] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Received event network-changed-6119ef11-a392-4c05-8c71-6e5b25f3cdc9 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 828.273311] env[61987]: DEBUG nova.compute.manager [req-db02cab5-8880-4fea-b11c-8feeb401b63c req-81d71756-7152-4601-b4db-9a4d6504332b service nova] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Refreshing instance network info cache due to event network-changed-6119ef11-a392-4c05-8c71-6e5b25f3cdc9. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 828.273472] env[61987]: DEBUG oslo_concurrency.lockutils [req-db02cab5-8880-4fea-b11c-8feeb401b63c req-81d71756-7152-4601-b4db-9a4d6504332b service nova] Acquiring lock "refresh_cache-6ed3f0e9-3c39-4406-9863-931a117961a8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.273620] env[61987]: DEBUG oslo_concurrency.lockutils [req-db02cab5-8880-4fea-b11c-8feeb401b63c req-81d71756-7152-4601-b4db-9a4d6504332b service nova] Acquired lock "refresh_cache-6ed3f0e9-3c39-4406-9863-931a117961a8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.273786] env[61987]: DEBUG nova.network.neutron [req-db02cab5-8880-4fea-b11c-8feeb401b63c req-81d71756-7152-4601-b4db-9a4d6504332b service nova] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Refreshing network info cache for port 6119ef11-a392-4c05-8c71-6e5b25f3cdc9 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 828.292282] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Created folder: Project (48d633a419cd468dae6134421c9c17fa) in parent group-v234219. [ 828.292593] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Creating folder: Instances. Parent ref: group-v234349. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 828.292767] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-267c618f-f6ed-484f-b3a4-e399eedc6d65 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.306074] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Created folder: Instances in parent group-v234349. [ 828.306392] env[61987]: DEBUG oslo.service.loopingcall [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.306611] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 828.306835] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4b77842e-4aa0-45bd-b369-a63f918898f3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.329153] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 828.329153] env[61987]: value = "task-1061723" [ 828.329153] env[61987]: _type = "Task" [ 828.329153] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.339355] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061723, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.400730] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061719, 'name': Rename_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.434723] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ec62fa3-5350-4b50-97dc-e34715d9544d tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Releasing lock "refresh_cache-10354ba0-5a64-4174-9c61-152d54a5dfcc" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.435234] env[61987]: DEBUG nova.compute.manager [None req-8ec62fa3-5350-4b50-97dc-e34715d9544d tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 828.435540] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8ec62fa3-5350-4b50-97dc-e34715d9544d tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 828.435917] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d117dff9-968c-4882-b7ce-71ac22f9c6f7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.447381] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdfab3a3-ac81-4ac3-924f-04e2c8360051 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.479991] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-8ec62fa3-5350-4b50-97dc-e34715d9544d tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 10354ba0-5a64-4174-9c61-152d54a5dfcc could not be found. [ 828.479991] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8ec62fa3-5350-4b50-97dc-e34715d9544d tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 828.479991] env[61987]: INFO nova.compute.manager [None req-8ec62fa3-5350-4b50-97dc-e34715d9544d tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Took 0.04 seconds to destroy the instance on the hypervisor. [ 828.480219] env[61987]: DEBUG oslo.service.loopingcall [None req-8ec62fa3-5350-4b50-97dc-e34715d9544d tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.480893] env[61987]: DEBUG nova.compute.manager [-] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 828.480893] env[61987]: DEBUG nova.network.neutron [-] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 828.498941] env[61987]: DEBUG nova.network.neutron [-] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 828.691114] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.513s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.691114] env[61987]: DEBUG nova.compute.manager [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 828.692571] env[61987]: DEBUG oslo_concurrency.lockutils [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.175s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.694166] env[61987]: INFO nova.compute.claims [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 828.706195] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061720, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.105168} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.706669] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 828.707618] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2667f5a4-8d5f-4901-bd4e-452d36783fd2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.726581] env[61987]: DEBUG nova.compute.manager [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 828.727012] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 828.742296] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 98d895ee-1d1f-49cd-a055-fe6c8462a692/98d895ee-1d1f-49cd-a055-fe6c8462a692.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 828.743869] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd94ada-c798-401c-87b8-e89714d1ea57 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.747338] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c7ad621-01c9-4696-832f-8f9e1a4e03d8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.775395] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 828.781122] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2dec19ac-2367-440a-b053-8acfbd85d8ba {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.782617] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Waiting for the task: (returnval){ [ 828.782617] env[61987]: value = "task-1061724" [ 828.782617] env[61987]: _type = "Task" [ 828.782617] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.783303] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061718, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.749566} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.783396] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 59d34772-26a5-4af2-82dc-da932e996745/59d34772-26a5-4af2-82dc-da932e996745.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 828.783599] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.787380] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fc83ce08-37be-4e28-92ae-3e1613a0cd8f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.789925] env[61987]: DEBUG oslo_vmware.api [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for the task: (returnval){ [ 828.789925] env[61987]: value = "task-1061725" [ 828.789925] env[61987]: _type = "Task" [ 828.789925] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.797385] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061724, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.797652] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for the task: (returnval){ [ 828.797652] env[61987]: value = "task-1061726" [ 828.797652] env[61987]: _type = "Task" [ 828.797652] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.804271] env[61987]: DEBUG oslo_vmware.api [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061725, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.808813] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061726, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.838630] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061723, 'name': CreateVM_Task, 'duration_secs': 0.4434} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.838779] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 828.840568] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.840568] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.840568] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 828.840568] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd0d9b8f-9994-49b0-a7d0-97985dd9f4f8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.844310] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Waiting for the task: (returnval){ [ 828.844310] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5200ec3e-d7a6-6ede-59a9-b64075f22d71" [ 828.844310] env[61987]: _type = "Task" [ 828.844310] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.854307] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5200ec3e-d7a6-6ede-59a9-b64075f22d71, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.902018] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061719, 'name': Rename_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.915992] env[61987]: DEBUG nova.compute.manager [req-cacd268d-5ef8-40b6-8819-62713ecf0aec req-fc609d87-050c-4b4b-9123-39acc0c97c3e service nova] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Received event network-vif-plugged-e88ac5c4-67fa-4013-ac31-ca59251955b5 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 828.916301] env[61987]: DEBUG oslo_concurrency.lockutils [req-cacd268d-5ef8-40b6-8819-62713ecf0aec req-fc609d87-050c-4b4b-9123-39acc0c97c3e service nova] Acquiring lock "056c181c-37da-49ac-ae1f-f7f2032c579a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.916461] env[61987]: DEBUG oslo_concurrency.lockutils [req-cacd268d-5ef8-40b6-8819-62713ecf0aec req-fc609d87-050c-4b4b-9123-39acc0c97c3e service nova] Lock "056c181c-37da-49ac-ae1f-f7f2032c579a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.916611] env[61987]: DEBUG oslo_concurrency.lockutils [req-cacd268d-5ef8-40b6-8819-62713ecf0aec req-fc609d87-050c-4b4b-9123-39acc0c97c3e service nova] Lock "056c181c-37da-49ac-ae1f-f7f2032c579a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.916781] env[61987]: DEBUG nova.compute.manager [req-cacd268d-5ef8-40b6-8819-62713ecf0aec req-fc609d87-050c-4b4b-9123-39acc0c97c3e service nova] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] No waiting events found dispatching network-vif-plugged-e88ac5c4-67fa-4013-ac31-ca59251955b5 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 828.921136] env[61987]: WARNING nova.compute.manager [req-cacd268d-5ef8-40b6-8819-62713ecf0aec req-fc609d87-050c-4b4b-9123-39acc0c97c3e service nova] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Received unexpected event network-vif-plugged-e88ac5c4-67fa-4013-ac31-ca59251955b5 for instance with vm_state building and task_state spawning. [ 828.921136] env[61987]: DEBUG nova.network.neutron [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Successfully updated port: e88ac5c4-67fa-4013-ac31-ca59251955b5 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 829.001107] env[61987]: DEBUG nova.network.neutron [-] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.095483] env[61987]: DEBUG nova.network.neutron [req-db02cab5-8880-4fea-b11c-8feeb401b63c req-81d71756-7152-4601-b4db-9a4d6504332b service nova] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Updated VIF entry in instance network info cache for port 6119ef11-a392-4c05-8c71-6e5b25f3cdc9. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 829.095877] env[61987]: DEBUG nova.network.neutron [req-db02cab5-8880-4fea-b11c-8feeb401b63c req-81d71756-7152-4601-b4db-9a4d6504332b service nova] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Updating instance_info_cache with network_info: [{"id": "6119ef11-a392-4c05-8c71-6e5b25f3cdc9", "address": "fa:16:3e:cc:5c:78", "network": {"id": "a91d7bba-6d0b-4ee1-a490-ea769b3b83ca", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-546229287-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "48d633a419cd468dae6134421c9c17fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "72781990-3cb3-42eb-9eb1-4040dedbf66f", "external-id": "cl2-zone-812", "segmentation_id": 812, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6119ef11-a3", "ovs_interfaceid": "6119ef11-a392-4c05-8c71-6e5b25f3cdc9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.201761] env[61987]: DEBUG nova.compute.utils [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 829.203313] env[61987]: DEBUG nova.compute.manager [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 829.203313] env[61987]: DEBUG nova.network.neutron [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 829.268416] env[61987]: DEBUG nova.policy [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e16f557dfdb04406b3fc772b891f05a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '22e62741abfb4471bbc7e11ae7d407ca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 829.293925] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061724, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.303761] env[61987]: DEBUG oslo_vmware.api [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061725, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.308150] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061726, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.448663} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.308375] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 829.309119] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbad435a-e558-4a34-b1f5-88b59e20cf4c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.328485] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 59d34772-26a5-4af2-82dc-da932e996745/59d34772-26a5-4af2-82dc-da932e996745.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 829.328775] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-15a0ffa9-f4c6-4581-b0d7-d8c43f51d7f4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.349585] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for the task: (returnval){ [ 829.349585] env[61987]: value = "task-1061727" [ 829.349585] env[61987]: _type = "Task" [ 829.349585] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.356111] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5200ec3e-d7a6-6ede-59a9-b64075f22d71, 'name': SearchDatastore_Task, 'duration_secs': 0.010188} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.356715] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.356950] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 829.357207] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.357363] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.357656] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 829.357788] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d812de94-2bdf-4741-99af-cce7ff0e06e7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.362439] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061727, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.368963] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 829.369089] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 829.369788] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6390d5d3-dffe-49b3-9470-45f3c29b298b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.374898] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Waiting for the task: (returnval){ [ 829.374898] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52bd1672-c730-d84e-7f65-dfa5a3cab39b" [ 829.374898] env[61987]: _type = "Task" [ 829.374898] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.383014] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52bd1672-c730-d84e-7f65-dfa5a3cab39b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.400540] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061719, 'name': Rename_Task, 'duration_secs': 1.215333} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.400845] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 829.401115] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-315a9bb3-418f-46fc-8c21-af8c8e65245b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.407234] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 829.407234] env[61987]: value = "task-1061728" [ 829.407234] env[61987]: _type = "Task" [ 829.407234] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.415390] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061728, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.421087] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "refresh_cache-056c181c-37da-49ac-ae1f-f7f2032c579a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.421319] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired lock "refresh_cache-056c181c-37da-49ac-ae1f-f7f2032c579a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.421517] env[61987]: DEBUG nova.network.neutron [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 829.505672] env[61987]: INFO nova.compute.manager [-] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Took 1.02 seconds to deallocate network for instance. [ 829.598939] env[61987]: DEBUG oslo_concurrency.lockutils [req-db02cab5-8880-4fea-b11c-8feeb401b63c req-81d71756-7152-4601-b4db-9a4d6504332b service nova] Releasing lock "refresh_cache-6ed3f0e9-3c39-4406-9863-931a117961a8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.599307] env[61987]: DEBUG nova.compute.manager [req-db02cab5-8880-4fea-b11c-8feeb401b63c req-81d71756-7152-4601-b4db-9a4d6504332b service nova] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Received event network-changed-b4c377b2-8e62-4cde-ac10-05c4cd2bde7d {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 829.599498] env[61987]: DEBUG nova.compute.manager [req-db02cab5-8880-4fea-b11c-8feeb401b63c req-81d71756-7152-4601-b4db-9a4d6504332b service nova] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Refreshing instance network info cache due to event network-changed-b4c377b2-8e62-4cde-ac10-05c4cd2bde7d. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 829.599747] env[61987]: DEBUG oslo_concurrency.lockutils [req-db02cab5-8880-4fea-b11c-8feeb401b63c req-81d71756-7152-4601-b4db-9a4d6504332b service nova] Acquiring lock "refresh_cache-fe22fc1b-61fb-4088-b5fe-b84eb4ec4099" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.599951] env[61987]: DEBUG oslo_concurrency.lockutils [req-db02cab5-8880-4fea-b11c-8feeb401b63c req-81d71756-7152-4601-b4db-9a4d6504332b service nova] Acquired lock "refresh_cache-fe22fc1b-61fb-4088-b5fe-b84eb4ec4099" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.600141] env[61987]: DEBUG nova.network.neutron [req-db02cab5-8880-4fea-b11c-8feeb401b63c req-81d71756-7152-4601-b4db-9a4d6504332b service nova] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Refreshing network info cache for port b4c377b2-8e62-4cde-ac10-05c4cd2bde7d {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 829.671196] env[61987]: DEBUG nova.network.neutron [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Successfully created port: 6d65f8d7-1b65-4aa2-93f5-60b85af049c1 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 829.706777] env[61987]: DEBUG nova.compute.manager [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 829.796127] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061724, 'name': ReconfigVM_Task, 'duration_secs': 0.900198} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.799333] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 98d895ee-1d1f-49cd-a055-fe6c8462a692/98d895ee-1d1f-49cd-a055-fe6c8462a692.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 829.802229] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0b9a3a87-8b3a-4527-a124-160ec0c17553 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.809042] env[61987]: DEBUG oslo_vmware.api [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061725, 'name': PowerOffVM_Task, 'duration_secs': 0.655052} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.810522] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 829.810522] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 829.810948] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Waiting for the task: (returnval){ [ 829.810948] env[61987]: value = "task-1061729" [ 829.810948] env[61987]: _type = "Task" [ 829.810948] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.811170] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8b142f89-cc2c-47c2-939b-29ea73f5c97f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.823318] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061729, 'name': Rename_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.861700] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061727, 'name': ReconfigVM_Task, 'duration_secs': 0.293505} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.863182] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 59d34772-26a5-4af2-82dc-da932e996745/59d34772-26a5-4af2-82dc-da932e996745.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 829.863896] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-46b8676a-664f-47cd-8ead-4e87c4a2275e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.870165] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for the task: (returnval){ [ 829.870165] env[61987]: value = "task-1061731" [ 829.870165] env[61987]: _type = "Task" [ 829.870165] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.882630] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061731, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.888696] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52bd1672-c730-d84e-7f65-dfa5a3cab39b, 'name': SearchDatastore_Task, 'duration_secs': 0.01541} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.889534] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2355ba79-f903-4e75-9a88-533c192ff48f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.898210] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Waiting for the task: (returnval){ [ 829.898210] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5247dd03-83a6-e42d-5ca7-8b977a6fd4ce" [ 829.898210] env[61987]: _type = "Task" [ 829.898210] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.906679] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5247dd03-83a6-e42d-5ca7-8b977a6fd4ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.910661] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 829.910875] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 829.911082] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Deleting the datastore file [datastore2] 4edd0159-6db1-41bd-a616-31a34e984059 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 829.914365] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5572375b-91a4-41c0-b295-75c85a194019 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.921855] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061728, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.923400] env[61987]: DEBUG oslo_vmware.api [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for the task: (returnval){ [ 829.923400] env[61987]: value = "task-1061732" [ 829.923400] env[61987]: _type = "Task" [ 829.923400] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.935731] env[61987]: DEBUG oslo_vmware.api [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061732, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.971808] env[61987]: DEBUG nova.network.neutron [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 830.013045] env[61987]: INFO nova.compute.manager [None req-8ec62fa3-5350-4b50-97dc-e34715d9544d tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Instance disappeared during terminate [ 830.013045] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ec62fa3-5350-4b50-97dc-e34715d9544d tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "10354ba0-5a64-4174-9c61-152d54a5dfcc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.254s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.038575] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f29a8cb3-bc35-44ea-886e-d118720b6585 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.047666] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f36573b-0d17-4b96-b838-16730b199aba {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.084115] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18c22aac-821f-43c1-9bf3-812111fde9fb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.092504] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc3c8aa8-142a-4f32-8897-ce304b94e5cb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.109718] env[61987]: DEBUG nova.compute.provider_tree [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.255166] env[61987]: DEBUG nova.network.neutron [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Updating instance_info_cache with network_info: [{"id": "e88ac5c4-67fa-4013-ac31-ca59251955b5", "address": "fa:16:3e:14:12:b9", "network": {"id": "ca68bf96-3f18-45ee-80fb-64a1260ed996", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-733595377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd47eb44bc334bf3ae5813905903ecbc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a27fd90b-16a5-43af-bede-ae36762ece00", "external-id": "nsx-vlan-transportzone-197", "segmentation_id": 197, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape88ac5c4-67", "ovs_interfaceid": "e88ac5c4-67fa-4013-ac31-ca59251955b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.323803] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061729, 'name': Rename_Task, 'duration_secs': 0.246446} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.324162] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 830.324430] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1d23e502-6eec-49b9-b6d4-effbba462200 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.330629] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Waiting for the task: (returnval){ [ 830.330629] env[61987]: value = "task-1061733" [ 830.330629] env[61987]: _type = "Task" [ 830.330629] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.340850] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061733, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.380658] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061731, 'name': Rename_Task, 'duration_secs': 0.140501} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.380972] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 830.381286] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-227bd9a0-d379-42cb-af64-8af029750868 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.390597] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for the task: (returnval){ [ 830.390597] env[61987]: value = "task-1061734" [ 830.390597] env[61987]: _type = "Task" [ 830.390597] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.398589] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061734, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.408279] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5247dd03-83a6-e42d-5ca7-8b977a6fd4ce, 'name': SearchDatastore_Task, 'duration_secs': 0.010986} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.408554] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.408813] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 6ed3f0e9-3c39-4406-9863-931a117961a8/6ed3f0e9-3c39-4406-9863-931a117961a8.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 830.409093] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-868cd570-12ff-4657-ac52-0699131b945e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.419013] env[61987]: DEBUG oslo_vmware.api [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061728, 'name': PowerOnVM_Task, 'duration_secs': 0.946647} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.422469] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 830.422802] env[61987]: INFO nova.compute.manager [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Took 9.14 seconds to spawn the instance on the hypervisor. [ 830.423029] env[61987]: DEBUG nova.compute.manager [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 830.423373] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Waiting for the task: (returnval){ [ 830.423373] env[61987]: value = "task-1061735" [ 830.423373] env[61987]: _type = "Task" [ 830.423373] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.424142] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a300e3-f3cb-476b-8e8d-f6386b7ab89c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.446130] env[61987]: DEBUG oslo_vmware.api [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Task: {'id': task-1061732, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141851} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.446575] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': task-1061735, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.447107] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 830.447418] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 830.447506] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 830.447667] env[61987]: INFO nova.compute.manager [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Took 1.72 seconds to destroy the instance on the hypervisor. [ 830.447939] env[61987]: DEBUG oslo.service.loopingcall [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.448175] env[61987]: DEBUG nova.compute.manager [-] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 830.448303] env[61987]: DEBUG nova.network.neutron [-] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 830.529184] env[61987]: DEBUG nova.network.neutron [req-db02cab5-8880-4fea-b11c-8feeb401b63c req-81d71756-7152-4601-b4db-9a4d6504332b service nova] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Updated VIF entry in instance network info cache for port b4c377b2-8e62-4cde-ac10-05c4cd2bde7d. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 830.529584] env[61987]: DEBUG nova.network.neutron [req-db02cab5-8880-4fea-b11c-8feeb401b63c req-81d71756-7152-4601-b4db-9a4d6504332b service nova] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Updating instance_info_cache with network_info: [{"id": "b4c377b2-8e62-4cde-ac10-05c4cd2bde7d", "address": "fa:16:3e:6f:ef:57", "network": {"id": "2bc28674-9598-4875-b1bc-4265efac93fe", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1333586968-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e802bb0ab61a41a79664c7e12fa0e2bf", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dad4f433-bb0b-45c7-8040-972ef2277f75", "external-id": "nsx-vlan-transportzone-451", "segmentation_id": 451, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4c377b2-8e", "ovs_interfaceid": "b4c377b2-8e62-4cde-ac10-05c4cd2bde7d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.613124] env[61987]: DEBUG nova.scheduler.client.report [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 830.718497] env[61987]: DEBUG nova.compute.manager [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 830.756127] env[61987]: DEBUG nova.virt.hardware [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 830.756487] env[61987]: DEBUG nova.virt.hardware [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 830.756715] env[61987]: DEBUG nova.virt.hardware [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 830.756930] env[61987]: DEBUG nova.virt.hardware [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 830.757578] env[61987]: DEBUG nova.virt.hardware [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 830.757788] env[61987]: DEBUG nova.virt.hardware [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 830.758175] env[61987]: DEBUG nova.virt.hardware [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 830.758346] env[61987]: DEBUG nova.virt.hardware [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 830.758561] env[61987]: DEBUG nova.virt.hardware [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 830.758803] env[61987]: DEBUG nova.virt.hardware [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 830.759075] env[61987]: DEBUG nova.virt.hardware [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 830.759666] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Releasing lock "refresh_cache-056c181c-37da-49ac-ae1f-f7f2032c579a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.760212] env[61987]: DEBUG nova.compute.manager [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Instance network_info: |[{"id": "e88ac5c4-67fa-4013-ac31-ca59251955b5", "address": "fa:16:3e:14:12:b9", "network": {"id": "ca68bf96-3f18-45ee-80fb-64a1260ed996", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-733595377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd47eb44bc334bf3ae5813905903ecbc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a27fd90b-16a5-43af-bede-ae36762ece00", "external-id": "nsx-vlan-transportzone-197", "segmentation_id": 197, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape88ac5c4-67", "ovs_interfaceid": "e88ac5c4-67fa-4013-ac31-ca59251955b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 830.761294] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87af9f00-25da-4544-9cd2-8ecae0214b50 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.765775] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:12:b9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a27fd90b-16a5-43af-bede-ae36762ece00', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e88ac5c4-67fa-4013-ac31-ca59251955b5', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 830.774231] env[61987]: DEBUG oslo.service.loopingcall [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.774982] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 830.775831] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b8534146-3172-41d8-bf81-22ab86e5ff54 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.800823] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-920d560e-5a3e-4790-84ff-6f4752c143b9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.819249] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 830.819249] env[61987]: value = "task-1061736" [ 830.819249] env[61987]: _type = "Task" [ 830.819249] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.844214] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061736, 'name': CreateVM_Task} progress is 15%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.849901] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061733, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.903887] env[61987]: DEBUG oslo_vmware.api [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061734, 'name': PowerOnVM_Task, 'duration_secs': 0.496846} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.904330] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 830.904614] env[61987]: DEBUG nova.compute.manager [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 830.905594] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd12f6c1-2bb8-411f-ada4-2c7bed1c4e23 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.938044] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': task-1061735, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.955462] env[61987]: INFO nova.compute.manager [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Took 50.97 seconds to build instance. [ 831.032865] env[61987]: DEBUG oslo_concurrency.lockutils [req-db02cab5-8880-4fea-b11c-8feeb401b63c req-81d71756-7152-4601-b4db-9a4d6504332b service nova] Releasing lock "refresh_cache-fe22fc1b-61fb-4088-b5fe-b84eb4ec4099" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.062944] env[61987]: DEBUG nova.compute.manager [req-dae4d901-4adc-49d4-8c91-4d59fd5e6386 req-b8b1d3af-0e61-44d8-9893-eb0acd2b3ef1 service nova] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Received event network-changed-e88ac5c4-67fa-4013-ac31-ca59251955b5 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 831.063167] env[61987]: DEBUG nova.compute.manager [req-dae4d901-4adc-49d4-8c91-4d59fd5e6386 req-b8b1d3af-0e61-44d8-9893-eb0acd2b3ef1 service nova] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Refreshing instance network info cache due to event network-changed-e88ac5c4-67fa-4013-ac31-ca59251955b5. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 831.063458] env[61987]: DEBUG oslo_concurrency.lockutils [req-dae4d901-4adc-49d4-8c91-4d59fd5e6386 req-b8b1d3af-0e61-44d8-9893-eb0acd2b3ef1 service nova] Acquiring lock "refresh_cache-056c181c-37da-49ac-ae1f-f7f2032c579a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.063666] env[61987]: DEBUG oslo_concurrency.lockutils [req-dae4d901-4adc-49d4-8c91-4d59fd5e6386 req-b8b1d3af-0e61-44d8-9893-eb0acd2b3ef1 service nova] Acquired lock "refresh_cache-056c181c-37da-49ac-ae1f-f7f2032c579a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.063815] env[61987]: DEBUG nova.network.neutron [req-dae4d901-4adc-49d4-8c91-4d59fd5e6386 req-b8b1d3af-0e61-44d8-9893-eb0acd2b3ef1 service nova] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Refreshing network info cache for port e88ac5c4-67fa-4013-ac31-ca59251955b5 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 831.118162] env[61987]: DEBUG oslo_concurrency.lockutils [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.118956] env[61987]: DEBUG nova.compute.manager [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 831.121487] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0522c70a-4abf-4f98-b1b3-b522ef661f68 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 19.279s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.328587] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061736, 'name': CreateVM_Task, 'duration_secs': 0.433426} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.328760] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 831.329502] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.329711] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.330043] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 831.330351] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98657659-cf6b-4bd1-ac16-411a25506c47 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.338479] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 831.338479] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]521e0b01-1e24-b9dc-e686-3afacbce9d47" [ 831.338479] env[61987]: _type = "Task" [ 831.338479] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.345588] env[61987]: DEBUG nova.network.neutron [-] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.346981] env[61987]: DEBUG oslo_vmware.api [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061733, 'name': PowerOnVM_Task, 'duration_secs': 0.648761} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.347822] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 831.347978] env[61987]: INFO nova.compute.manager [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Took 8.11 seconds to spawn the instance on the hypervisor. [ 831.348183] env[61987]: DEBUG nova.compute.manager [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 831.349161] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7100724f-b6ae-4bf2-ba4d-33ac9eb851ed {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.355376] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]521e0b01-1e24-b9dc-e686-3afacbce9d47, 'name': SearchDatastore_Task, 'duration_secs': 0.009281} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.356050] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.356560] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 831.356560] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.356699] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.356852] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 831.357133] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-484d8897-f844-4c48-8fba-be55e716eb39 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.369444] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 831.369633] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 831.371402] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa02a9b7-0f1f-4a7c-9348-679efd0120ef {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.375745] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 831.375745] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]520d4feb-0ad7-1921-b99e-5ffaf77df466" [ 831.375745] env[61987]: _type = "Task" [ 831.375745] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.385230] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]520d4feb-0ad7-1921-b99e-5ffaf77df466, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.425497] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.438904] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': task-1061735, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518658} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.442019] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 6ed3f0e9-3c39-4406-9863-931a117961a8/6ed3f0e9-3c39-4406-9863-931a117961a8.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 831.442019] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 831.442019] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a9c9e11-fca1-478d-ab78-75f3040d99b6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.446011] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Waiting for the task: (returnval){ [ 831.446011] env[61987]: value = "task-1061737" [ 831.446011] env[61987]: _type = "Task" [ 831.446011] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.453848] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': task-1061737, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.454883] env[61987]: DEBUG nova.network.neutron [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Successfully updated port: 6d65f8d7-1b65-4aa2-93f5-60b85af049c1 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 831.457629] env[61987]: DEBUG oslo_concurrency.lockutils [None req-77c37f22-1d30-4955-be53-8d3d4e5a93c8 tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Lock "93706a54-1fc4-4aed-8807-ec2c3d3694b4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.482s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.596594] env[61987]: DEBUG nova.compute.manager [req-ab9fd478-e9e5-4ac4-968b-894d1134850d req-3c50333d-ddcd-4bf3-903d-d4f58072464d service nova] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Received event network-vif-plugged-6d65f8d7-1b65-4aa2-93f5-60b85af049c1 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 831.596935] env[61987]: DEBUG oslo_concurrency.lockutils [req-ab9fd478-e9e5-4ac4-968b-894d1134850d req-3c50333d-ddcd-4bf3-903d-d4f58072464d service nova] Acquiring lock "1961c2c7-4400-4559-9f7c-bec8e8e263b0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.597173] env[61987]: DEBUG oslo_concurrency.lockutils [req-ab9fd478-e9e5-4ac4-968b-894d1134850d req-3c50333d-ddcd-4bf3-903d-d4f58072464d service nova] Lock "1961c2c7-4400-4559-9f7c-bec8e8e263b0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.597433] env[61987]: DEBUG oslo_concurrency.lockutils [req-ab9fd478-e9e5-4ac4-968b-894d1134850d req-3c50333d-ddcd-4bf3-903d-d4f58072464d service nova] Lock "1961c2c7-4400-4559-9f7c-bec8e8e263b0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.597636] env[61987]: DEBUG nova.compute.manager [req-ab9fd478-e9e5-4ac4-968b-894d1134850d req-3c50333d-ddcd-4bf3-903d-d4f58072464d service nova] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] No waiting events found dispatching network-vif-plugged-6d65f8d7-1b65-4aa2-93f5-60b85af049c1 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 831.597925] env[61987]: WARNING nova.compute.manager [req-ab9fd478-e9e5-4ac4-968b-894d1134850d req-3c50333d-ddcd-4bf3-903d-d4f58072464d service nova] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Received unexpected event network-vif-plugged-6d65f8d7-1b65-4aa2-93f5-60b85af049c1 for instance with vm_state building and task_state spawning. [ 831.624897] env[61987]: DEBUG nova.compute.utils [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 831.630230] env[61987]: DEBUG nova.compute.manager [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 831.631450] env[61987]: DEBUG nova.network.neutron [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 831.712615] env[61987]: DEBUG nova.policy [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3011f049be54b2cb8922d646b94a310', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd828262de7b14d2ab38193d7d34e8f7a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 831.843033] env[61987]: DEBUG nova.network.neutron [req-dae4d901-4adc-49d4-8c91-4d59fd5e6386 req-b8b1d3af-0e61-44d8-9893-eb0acd2b3ef1 service nova] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Updated VIF entry in instance network info cache for port e88ac5c4-67fa-4013-ac31-ca59251955b5. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 831.843489] env[61987]: DEBUG nova.network.neutron [req-dae4d901-4adc-49d4-8c91-4d59fd5e6386 req-b8b1d3af-0e61-44d8-9893-eb0acd2b3ef1 service nova] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Updating instance_info_cache with network_info: [{"id": "e88ac5c4-67fa-4013-ac31-ca59251955b5", "address": "fa:16:3e:14:12:b9", "network": {"id": "ca68bf96-3f18-45ee-80fb-64a1260ed996", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-733595377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd47eb44bc334bf3ae5813905903ecbc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a27fd90b-16a5-43af-bede-ae36762ece00", "external-id": "nsx-vlan-transportzone-197", "segmentation_id": 197, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape88ac5c4-67", "ovs_interfaceid": "e88ac5c4-67fa-4013-ac31-ca59251955b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.847922] env[61987]: INFO nova.compute.manager [-] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Took 1.40 seconds to deallocate network for instance. [ 831.873327] env[61987]: INFO nova.compute.manager [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Took 50.44 seconds to build instance. [ 831.889047] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]520d4feb-0ad7-1921-b99e-5ffaf77df466, 'name': SearchDatastore_Task, 'duration_secs': 0.008152} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.889899] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6555858e-891f-4f03-becb-d8e9c7a23e4b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.896028] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 831.896028] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52124249-1190-20e6-a2ae-55d15e8c8d93" [ 831.896028] env[61987]: _type = "Task" [ 831.896028] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.908938] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52124249-1190-20e6-a2ae-55d15e8c8d93, 'name': SearchDatastore_Task, 'duration_secs': 0.009274} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.909372] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.909752] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 056c181c-37da-49ac-ae1f-f7f2032c579a/056c181c-37da-49ac-ae1f-f7f2032c579a.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 831.910083] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e631f56a-508d-4488-b93d-dfc4f7caf0f9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.919033] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 831.919033] env[61987]: value = "task-1061738" [ 831.919033] env[61987]: _type = "Task" [ 831.919033] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.927324] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061738, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.956506] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': task-1061737, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072098} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.959521] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "refresh_cache-1961c2c7-4400-4559-9f7c-bec8e8e263b0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.959909] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired lock "refresh_cache-1961c2c7-4400-4559-9f7c-bec8e8e263b0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.959909] env[61987]: DEBUG nova.network.neutron [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 831.963603] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 831.963603] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12abee08-66aa-4b6a-b7ef-61e55bb4e870 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.987691] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 6ed3f0e9-3c39-4406-9863-931a117961a8/6ed3f0e9-3c39-4406-9863-931a117961a8.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 831.991092] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-237407be-6bdc-496b-bcfb-74a71eb04565 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.016586] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Waiting for the task: (returnval){ [ 832.016586] env[61987]: value = "task-1061739" [ 832.016586] env[61987]: _type = "Task" [ 832.016586] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.022656] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': task-1061739, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.028946] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cf0988a-c292-4866-8053-33eae5ad3e25 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.036520] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ecacdff-aada-4125-9613-a23dd5a10381 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.067436] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4c782fc-6197-4a67-b3dd-1a2589db979a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.077269] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20b9eedf-d3df-46ac-934f-982da7ff3f4d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.089378] env[61987]: DEBUG nova.compute.provider_tree [None req-0522c70a-4abf-4f98-b1b3-b522ef661f68 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.131230] env[61987]: DEBUG nova.compute.manager [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 832.258663] env[61987]: DEBUG nova.network.neutron [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Successfully created port: ac11f37b-72cc-4fa6-a61a-4535486b9d7f {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 832.351307] env[61987]: DEBUG oslo_concurrency.lockutils [req-dae4d901-4adc-49d4-8c91-4d59fd5e6386 req-b8b1d3af-0e61-44d8-9893-eb0acd2b3ef1 service nova] Releasing lock "refresh_cache-056c181c-37da-49ac-ae1f-f7f2032c579a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.351528] env[61987]: DEBUG nova.compute.manager [req-dae4d901-4adc-49d4-8c91-4d59fd5e6386 req-b8b1d3af-0e61-44d8-9893-eb0acd2b3ef1 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Received event network-vif-deleted-39309545-c724-44c7-932e-7784c9e31fa7 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 832.351718] env[61987]: INFO nova.compute.manager [req-dae4d901-4adc-49d4-8c91-4d59fd5e6386 req-b8b1d3af-0e61-44d8-9893-eb0acd2b3ef1 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Neutron deleted interface 39309545-c724-44c7-932e-7784c9e31fa7; detaching it from the instance and deleting it from the info cache [ 832.351902] env[61987]: DEBUG nova.network.neutron [req-dae4d901-4adc-49d4-8c91-4d59fd5e6386 req-b8b1d3af-0e61-44d8-9893-eb0acd2b3ef1 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.358036] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.375863] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4c5a77f2-1a9b-41d4-b3be-680c7ead80ad tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Lock "98d895ee-1d1f-49cd-a055-fe6c8462a692" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.955s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.439242] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061738, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.439669] env[61987]: DEBUG oslo_concurrency.lockutils [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Acquiring lock "59d34772-26a5-4af2-82dc-da932e996745" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.439907] env[61987]: DEBUG oslo_concurrency.lockutils [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Lock "59d34772-26a5-4af2-82dc-da932e996745" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.440132] env[61987]: DEBUG oslo_concurrency.lockutils [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Acquiring lock "59d34772-26a5-4af2-82dc-da932e996745-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.440336] env[61987]: DEBUG oslo_concurrency.lockutils [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Lock "59d34772-26a5-4af2-82dc-da932e996745-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.440513] env[61987]: DEBUG oslo_concurrency.lockutils [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Lock "59d34772-26a5-4af2-82dc-da932e996745-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.443599] env[61987]: INFO nova.compute.manager [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Terminating instance [ 832.528373] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': task-1061739, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.528373] env[61987]: DEBUG nova.network.neutron [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 832.593333] env[61987]: DEBUG nova.scheduler.client.report [None req-0522c70a-4abf-4f98-b1b3-b522ef661f68 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 832.753851] env[61987]: DEBUG nova.network.neutron [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Updating instance_info_cache with network_info: [{"id": "6d65f8d7-1b65-4aa2-93f5-60b85af049c1", "address": "fa:16:3e:8e:fd:4a", "network": {"id": "d5ed1e9e-9033-441e-bfcb-2ac643d7f03f", "bridge": "br-int", "label": "tempest-ServersTestJSON-8466993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e62741abfb4471bbc7e11ae7d407ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39ab9baf-90cd-4fe2-8d56-434f8210fc19", "external-id": "nsx-vlan-transportzone-713", "segmentation_id": 713, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d65f8d7-1b", "ovs_interfaceid": "6d65f8d7-1b65-4aa2-93f5-60b85af049c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.855640] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f13dac40-e36a-445d-be06-fcf61cab4fbe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.864838] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c04d49a0-41b1-4d48-8f4a-b107adfa0025 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.898650] env[61987]: DEBUG nova.compute.manager [req-dae4d901-4adc-49d4-8c91-4d59fd5e6386 req-b8b1d3af-0e61-44d8-9893-eb0acd2b3ef1 service nova] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Detach interface failed, port_id=39309545-c724-44c7-932e-7784c9e31fa7, reason: Instance 4edd0159-6db1-41bd-a616-31a34e984059 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 832.928665] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061738, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.557333} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.928946] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 056c181c-37da-49ac-ae1f-f7f2032c579a/056c181c-37da-49ac-ae1f-f7f2032c579a.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 832.929218] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 832.929489] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cbdbc1c0-f82d-4f91-8e2d-7e490d2b7ee0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.935661] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 832.935661] env[61987]: value = "task-1061740" [ 832.935661] env[61987]: _type = "Task" [ 832.935661] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.944055] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061740, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.949141] env[61987]: DEBUG oslo_concurrency.lockutils [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Acquiring lock "refresh_cache-59d34772-26a5-4af2-82dc-da932e996745" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.949322] env[61987]: DEBUG oslo_concurrency.lockutils [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Acquired lock "refresh_cache-59d34772-26a5-4af2-82dc-da932e996745" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.949511] env[61987]: DEBUG nova.network.neutron [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 833.025605] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': task-1061739, 'name': ReconfigVM_Task, 'duration_secs': 0.573759} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.025893] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 6ed3f0e9-3c39-4406-9863-931a117961a8/6ed3f0e9-3c39-4406-9863-931a117961a8.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 833.026558] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7645355b-9810-4f2d-bf8d-682a62e6c359 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.032537] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Waiting for the task: (returnval){ [ 833.032537] env[61987]: value = "task-1061741" [ 833.032537] env[61987]: _type = "Task" [ 833.032537] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.040674] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': task-1061741, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.142900] env[61987]: DEBUG nova.compute.manager [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 833.173744] env[61987]: DEBUG nova.virt.hardware [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='00460c3ff38118a0ee563b3dadc0df0c',container_format='bare',created_at=2024-09-30T23:11:51Z,direct_url=,disk_format='vmdk',id=92117f13-9c42-454c-87bb-8c7e90801e79,min_disk=1,min_ram=0,name='tempest-test-snap-2036877708',owner='d828262de7b14d2ab38193d7d34e8f7a',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-09-30T23:12:08Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 833.174015] env[61987]: DEBUG nova.virt.hardware [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 833.174248] env[61987]: DEBUG nova.virt.hardware [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 833.174458] env[61987]: DEBUG nova.virt.hardware [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 833.174616] env[61987]: DEBUG nova.virt.hardware [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 833.175197] env[61987]: DEBUG nova.virt.hardware [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 833.175197] env[61987]: DEBUG nova.virt.hardware [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 833.175309] env[61987]: DEBUG nova.virt.hardware [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 833.175436] env[61987]: DEBUG nova.virt.hardware [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 833.175609] env[61987]: DEBUG nova.virt.hardware [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 833.177026] env[61987]: DEBUG nova.virt.hardware [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 833.177026] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b722dc9-d592-4cbf-a315-fbac897515e6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.184782] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76f1c83f-82c2-4817-b414-c9434742c911 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.206032] env[61987]: DEBUG nova.compute.manager [req-490c663e-bac5-4d0b-a5d4-f6a4918f395b req-c22b6bb5-863d-436e-aeb4-6eb3a3cf2cca service nova] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Received event network-changed-6fcb037d-fd48-490d-9898-b952ffaad4ef {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 833.206120] env[61987]: DEBUG nova.compute.manager [req-490c663e-bac5-4d0b-a5d4-f6a4918f395b req-c22b6bb5-863d-436e-aeb4-6eb3a3cf2cca service nova] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Refreshing instance network info cache due to event network-changed-6fcb037d-fd48-490d-9898-b952ffaad4ef. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 833.206327] env[61987]: DEBUG oslo_concurrency.lockutils [req-490c663e-bac5-4d0b-a5d4-f6a4918f395b req-c22b6bb5-863d-436e-aeb4-6eb3a3cf2cca service nova] Acquiring lock "refresh_cache-93706a54-1fc4-4aed-8807-ec2c3d3694b4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.206513] env[61987]: DEBUG oslo_concurrency.lockutils [req-490c663e-bac5-4d0b-a5d4-f6a4918f395b req-c22b6bb5-863d-436e-aeb4-6eb3a3cf2cca service nova] Acquired lock "refresh_cache-93706a54-1fc4-4aed-8807-ec2c3d3694b4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.206648] env[61987]: DEBUG nova.network.neutron [req-490c663e-bac5-4d0b-a5d4-f6a4918f395b req-c22b6bb5-863d-436e-aeb4-6eb3a3cf2cca service nova] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Refreshing network info cache for port 6fcb037d-fd48-490d-9898-b952ffaad4ef {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 833.256743] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Releasing lock "refresh_cache-1961c2c7-4400-4559-9f7c-bec8e8e263b0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.256743] env[61987]: DEBUG nova.compute.manager [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Instance network_info: |[{"id": "6d65f8d7-1b65-4aa2-93f5-60b85af049c1", "address": "fa:16:3e:8e:fd:4a", "network": {"id": "d5ed1e9e-9033-441e-bfcb-2ac643d7f03f", "bridge": "br-int", "label": "tempest-ServersTestJSON-8466993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e62741abfb4471bbc7e11ae7d407ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39ab9baf-90cd-4fe2-8d56-434f8210fc19", "external-id": "nsx-vlan-transportzone-713", "segmentation_id": 713, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d65f8d7-1b", "ovs_interfaceid": "6d65f8d7-1b65-4aa2-93f5-60b85af049c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 833.257169] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:fd:4a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '39ab9baf-90cd-4fe2-8d56-434f8210fc19', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6d65f8d7-1b65-4aa2-93f5-60b85af049c1', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 833.265121] env[61987]: DEBUG oslo.service.loopingcall [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.265387] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 833.265625] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-19d618ad-fc20-4b64-9ce2-8b1ae0487e23 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.286801] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 833.286801] env[61987]: value = "task-1061742" [ 833.286801] env[61987]: _type = "Task" [ 833.286801] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.295249] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061742, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.380195] env[61987]: INFO nova.compute.manager [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Rescuing [ 833.380523] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Acquiring lock "refresh_cache-98d895ee-1d1f-49cd-a055-fe6c8462a692" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.380684] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Acquired lock "refresh_cache-98d895ee-1d1f-49cd-a055-fe6c8462a692" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.381053] env[61987]: DEBUG nova.network.neutron [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 833.446268] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061740, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088332} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.446640] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 833.447487] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6097ea4b-241b-4792-be9a-60e5f2ca3feb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.472171] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 056c181c-37da-49ac-ae1f-f7f2032c579a/056c181c-37da-49ac-ae1f-f7f2032c579a.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 833.472480] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a84b6f46-437b-4d3f-a229-6490b63578ca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.493648] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 833.493648] env[61987]: value = "task-1061743" [ 833.493648] env[61987]: _type = "Task" [ 833.493648] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.501697] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061743, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.502611] env[61987]: DEBUG nova.network.neutron [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 833.544207] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': task-1061741, 'name': Rename_Task, 'duration_secs': 0.160517} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.544710] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 833.545017] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eef252ba-6c3c-41d2-a8df-bc414c44f6a4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.551382] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Waiting for the task: (returnval){ [ 833.551382] env[61987]: value = "task-1061744" [ 833.551382] env[61987]: _type = "Task" [ 833.551382] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.560716] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': task-1061744, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.568305] env[61987]: DEBUG nova.network.neutron [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.603299] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0522c70a-4abf-4f98-b1b3-b522ef661f68 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.482s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.607179] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.672s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.609460] env[61987]: INFO nova.compute.claims [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 833.695520] env[61987]: DEBUG nova.compute.manager [req-06a6781c-2880-4d22-874e-fd8f8ce90f46 req-3726e9bc-0382-40ca-ae87-b784ea10202c service nova] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Received event network-changed-6d65f8d7-1b65-4aa2-93f5-60b85af049c1 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 833.695728] env[61987]: DEBUG nova.compute.manager [req-06a6781c-2880-4d22-874e-fd8f8ce90f46 req-3726e9bc-0382-40ca-ae87-b784ea10202c service nova] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Refreshing instance network info cache due to event network-changed-6d65f8d7-1b65-4aa2-93f5-60b85af049c1. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 833.695957] env[61987]: DEBUG oslo_concurrency.lockutils [req-06a6781c-2880-4d22-874e-fd8f8ce90f46 req-3726e9bc-0382-40ca-ae87-b784ea10202c service nova] Acquiring lock "refresh_cache-1961c2c7-4400-4559-9f7c-bec8e8e263b0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.696126] env[61987]: DEBUG oslo_concurrency.lockutils [req-06a6781c-2880-4d22-874e-fd8f8ce90f46 req-3726e9bc-0382-40ca-ae87-b784ea10202c service nova] Acquired lock "refresh_cache-1961c2c7-4400-4559-9f7c-bec8e8e263b0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.696307] env[61987]: DEBUG nova.network.neutron [req-06a6781c-2880-4d22-874e-fd8f8ce90f46 req-3726e9bc-0382-40ca-ae87-b784ea10202c service nova] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Refreshing network info cache for port 6d65f8d7-1b65-4aa2-93f5-60b85af049c1 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 833.797642] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061742, 'name': CreateVM_Task, 'duration_secs': 0.365035} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.797839] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 833.798652] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.798922] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.799298] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 833.799567] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07e2f63c-1dbf-478e-af84-382f701aaf0b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.809025] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 833.809025] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]526419c8-6213-2663-1d2e-0047bea037c2" [ 833.809025] env[61987]: _type = "Task" [ 833.809025] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.815832] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]526419c8-6213-2663-1d2e-0047bea037c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.000501] env[61987]: DEBUG nova.network.neutron [req-490c663e-bac5-4d0b-a5d4-f6a4918f395b req-c22b6bb5-863d-436e-aeb4-6eb3a3cf2cca service nova] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Updated VIF entry in instance network info cache for port 6fcb037d-fd48-490d-9898-b952ffaad4ef. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 834.000934] env[61987]: DEBUG nova.network.neutron [req-490c663e-bac5-4d0b-a5d4-f6a4918f395b req-c22b6bb5-863d-436e-aeb4-6eb3a3cf2cca service nova] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Updating instance_info_cache with network_info: [{"id": "6fcb037d-fd48-490d-9898-b952ffaad4ef", "address": "fa:16:3e:14:a7:3a", "network": {"id": "35c53c0f-af9d-4934-84c7-ea4c5252fa58", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-343092073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.133", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d32dc833164a428dad56d1eeb9a7c3eb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "790c811b-3fa6-49f8-87ac-c51450911137", "external-id": "nsx-vlan-transportzone-908", "segmentation_id": 908, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6fcb037d-fd", "ovs_interfaceid": "6fcb037d-fd48-490d-9898-b952ffaad4ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.005435] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061743, 'name': ReconfigVM_Task, 'duration_secs': 0.383386} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.009017] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 056c181c-37da-49ac-ae1f-f7f2032c579a/056c181c-37da-49ac-ae1f-f7f2032c579a.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 834.009017] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c65264f6-7a28-4c8c-92a8-143cec38606c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.013710] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 834.013710] env[61987]: value = "task-1061745" [ 834.013710] env[61987]: _type = "Task" [ 834.013710] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.025817] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061745, 'name': Rename_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.061154] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': task-1061744, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.071448] env[61987]: DEBUG oslo_concurrency.lockutils [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Releasing lock "refresh_cache-59d34772-26a5-4af2-82dc-da932e996745" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.071886] env[61987]: DEBUG nova.compute.manager [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 834.072110] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 834.072947] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92c7ed70-f9a4-439d-a670-578f599ff4fd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.080395] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 834.080395] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eeeaa5d2-62e9-40e8-a7e6-84d6f6e85417 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.085824] env[61987]: DEBUG oslo_vmware.api [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for the task: (returnval){ [ 834.085824] env[61987]: value = "task-1061746" [ 834.085824] env[61987]: _type = "Task" [ 834.085824] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.093638] env[61987]: DEBUG oslo_vmware.api [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061746, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.149855] env[61987]: DEBUG nova.network.neutron [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Successfully updated port: ac11f37b-72cc-4fa6-a61a-4535486b9d7f {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 834.185228] env[61987]: INFO nova.scheduler.client.report [None req-0522c70a-4abf-4f98-b1b3-b522ef661f68 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Deleted allocation for migration c7ecbdeb-124a-4889-b4e4-2db11acba530 [ 834.318561] env[61987]: DEBUG nova.network.neutron [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Updating instance_info_cache with network_info: [{"id": "69ce8bb3-dc28-422e-96a8-833062bd6b77", "address": "fa:16:3e:ab:10:c2", "network": {"id": "35162a8d-60d3-465b-9330-64bbd235d6a2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-981766067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8dffc4d03e514fb8932b80718375084b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69ce8bb3-dc", "ovs_interfaceid": "69ce8bb3-dc28-422e-96a8-833062bd6b77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.327594] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]526419c8-6213-2663-1d2e-0047bea037c2, 'name': SearchDatastore_Task, 'duration_secs': 0.020388} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.328014] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.328986] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 834.329377] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.329620] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.329901] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 834.333946] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f2af1c10-cd6e-4fd8-85e6-044606aa7771 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.344189] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 834.344521] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 834.346030] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d173c15d-7847-4ff9-bca0-ce7957f1a4aa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.354926] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 834.354926] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]523113c5-9bcc-7b52-287c-38bf3712bea8" [ 834.354926] env[61987]: _type = "Task" [ 834.354926] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.367276] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523113c5-9bcc-7b52-287c-38bf3712bea8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.506649] env[61987]: DEBUG oslo_concurrency.lockutils [req-490c663e-bac5-4d0b-a5d4-f6a4918f395b req-c22b6bb5-863d-436e-aeb4-6eb3a3cf2cca service nova] Releasing lock "refresh_cache-93706a54-1fc4-4aed-8807-ec2c3d3694b4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.527704] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061745, 'name': Rename_Task, 'duration_secs': 0.224955} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.528029] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 834.528766] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-651cf4c2-9628-4beb-84ea-ccbb6b72403c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.535649] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 834.535649] env[61987]: value = "task-1061747" [ 834.535649] env[61987]: _type = "Task" [ 834.535649] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.544649] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061747, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.545882] env[61987]: DEBUG nova.network.neutron [req-06a6781c-2880-4d22-874e-fd8f8ce90f46 req-3726e9bc-0382-40ca-ae87-b784ea10202c service nova] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Updated VIF entry in instance network info cache for port 6d65f8d7-1b65-4aa2-93f5-60b85af049c1. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 834.546314] env[61987]: DEBUG nova.network.neutron [req-06a6781c-2880-4d22-874e-fd8f8ce90f46 req-3726e9bc-0382-40ca-ae87-b784ea10202c service nova] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Updating instance_info_cache with network_info: [{"id": "6d65f8d7-1b65-4aa2-93f5-60b85af049c1", "address": "fa:16:3e:8e:fd:4a", "network": {"id": "d5ed1e9e-9033-441e-bfcb-2ac643d7f03f", "bridge": "br-int", "label": "tempest-ServersTestJSON-8466993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e62741abfb4471bbc7e11ae7d407ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39ab9baf-90cd-4fe2-8d56-434f8210fc19", "external-id": "nsx-vlan-transportzone-713", "segmentation_id": 713, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d65f8d7-1b", "ovs_interfaceid": "6d65f8d7-1b65-4aa2-93f5-60b85af049c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.563253] env[61987]: DEBUG oslo_vmware.api [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': task-1061744, 'name': PowerOnVM_Task, 'duration_secs': 0.723682} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.564079] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 834.564079] env[61987]: INFO nova.compute.manager [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Took 8.84 seconds to spawn the instance on the hypervisor. [ 834.564079] env[61987]: DEBUG nova.compute.manager [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 834.564927] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d76d93c0-2306-4417-aaf1-bfd20dc36180 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.597033] env[61987]: DEBUG oslo_vmware.api [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061746, 'name': PowerOffVM_Task, 'duration_secs': 0.172045} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.597033] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 834.597033] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 834.597278] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-01456585-1819-4fbc-9228-e02432e553f2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.625731] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 834.626062] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 834.626119] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Deleting the datastore file [datastore2] 59d34772-26a5-4af2-82dc-da932e996745 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 834.626691] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6803c623-40c1-4647-806a-96339f303b97 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.632260] env[61987]: DEBUG oslo_vmware.api [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for the task: (returnval){ [ 834.632260] env[61987]: value = "task-1061749" [ 834.632260] env[61987]: _type = "Task" [ 834.632260] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.640573] env[61987]: DEBUG oslo_vmware.api [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061749, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.653862] env[61987]: DEBUG oslo_concurrency.lockutils [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "refresh_cache-8f9aab3e-96ee-44c6-8d25-706a0ff857c4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.654018] env[61987]: DEBUG oslo_concurrency.lockutils [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquired lock "refresh_cache-8f9aab3e-96ee-44c6-8d25-706a0ff857c4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.654199] env[61987]: DEBUG nova.network.neutron [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 834.692041] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0522c70a-4abf-4f98-b1b3-b522ef661f68 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 25.664s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.821204] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Releasing lock "refresh_cache-98d895ee-1d1f-49cd-a055-fe6c8462a692" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.866687] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523113c5-9bcc-7b52-287c-38bf3712bea8, 'name': SearchDatastore_Task, 'duration_secs': 0.014162} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.870346] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-482fb5e8-68c5-4c25-a52e-8b1995204208 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.876356] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 834.876356] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]520c3c63-994a-a775-0130-a2abaf117242" [ 834.876356] env[61987]: _type = "Task" [ 834.876356] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.885786] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]520c3c63-994a-a775-0130-a2abaf117242, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.898457] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d66c02e-22bb-4f07-a721-aac3c8420327 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.906374] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dad1d314-52ff-420f-94f0-9f0e09bb6ab2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.938912] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56495a99-67e1-41b0-9b4b-df80006c3b95 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.946638] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48c620a7-354c-4556-906d-c24f43d94334 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.960273] env[61987]: DEBUG nova.compute.provider_tree [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.048953] env[61987]: DEBUG oslo_concurrency.lockutils [req-06a6781c-2880-4d22-874e-fd8f8ce90f46 req-3726e9bc-0382-40ca-ae87-b784ea10202c service nova] Releasing lock "refresh_cache-1961c2c7-4400-4559-9f7c-bec8e8e263b0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.051836] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061747, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.088314] env[61987]: INFO nova.compute.manager [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Took 44.95 seconds to build instance. [ 835.142868] env[61987]: DEBUG oslo_vmware.api [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Task: {'id': task-1061749, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.262995} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.143189] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 835.143389] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 835.143572] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 835.143752] env[61987]: INFO nova.compute.manager [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Took 1.07 seconds to destroy the instance on the hypervisor. [ 835.144052] env[61987]: DEBUG oslo.service.loopingcall [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 835.144243] env[61987]: DEBUG nova.compute.manager [-] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 835.144346] env[61987]: DEBUG nova.network.neutron [-] [instance: 59d34772-26a5-4af2-82dc-da932e996745] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 835.164949] env[61987]: DEBUG nova.network.neutron [-] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 835.189945] env[61987]: DEBUG nova.network.neutron [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 835.336724] env[61987]: DEBUG nova.compute.manager [req-2ba5f0d3-40f2-476d-8fe4-bbc840df40da req-4a62f380-4bf9-43a7-9855-38bfb40c7934 service nova] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Received event network-vif-plugged-ac11f37b-72cc-4fa6-a61a-4535486b9d7f {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 835.337086] env[61987]: DEBUG oslo_concurrency.lockutils [req-2ba5f0d3-40f2-476d-8fe4-bbc840df40da req-4a62f380-4bf9-43a7-9855-38bfb40c7934 service nova] Acquiring lock "8f9aab3e-96ee-44c6-8d25-706a0ff857c4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.337417] env[61987]: DEBUG oslo_concurrency.lockutils [req-2ba5f0d3-40f2-476d-8fe4-bbc840df40da req-4a62f380-4bf9-43a7-9855-38bfb40c7934 service nova] Lock "8f9aab3e-96ee-44c6-8d25-706a0ff857c4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.337742] env[61987]: DEBUG oslo_concurrency.lockutils [req-2ba5f0d3-40f2-476d-8fe4-bbc840df40da req-4a62f380-4bf9-43a7-9855-38bfb40c7934 service nova] Lock "8f9aab3e-96ee-44c6-8d25-706a0ff857c4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.337991] env[61987]: DEBUG nova.compute.manager [req-2ba5f0d3-40f2-476d-8fe4-bbc840df40da req-4a62f380-4bf9-43a7-9855-38bfb40c7934 service nova] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] No waiting events found dispatching network-vif-plugged-ac11f37b-72cc-4fa6-a61a-4535486b9d7f {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 835.338254] env[61987]: WARNING nova.compute.manager [req-2ba5f0d3-40f2-476d-8fe4-bbc840df40da req-4a62f380-4bf9-43a7-9855-38bfb40c7934 service nova] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Received unexpected event network-vif-plugged-ac11f37b-72cc-4fa6-a61a-4535486b9d7f for instance with vm_state building and task_state spawning. [ 835.338988] env[61987]: DEBUG nova.compute.manager [req-2ba5f0d3-40f2-476d-8fe4-bbc840df40da req-4a62f380-4bf9-43a7-9855-38bfb40c7934 service nova] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Received event network-changed-ac11f37b-72cc-4fa6-a61a-4535486b9d7f {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 835.338988] env[61987]: DEBUG nova.compute.manager [req-2ba5f0d3-40f2-476d-8fe4-bbc840df40da req-4a62f380-4bf9-43a7-9855-38bfb40c7934 service nova] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Refreshing instance network info cache due to event network-changed-ac11f37b-72cc-4fa6-a61a-4535486b9d7f. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 835.338988] env[61987]: DEBUG oslo_concurrency.lockutils [req-2ba5f0d3-40f2-476d-8fe4-bbc840df40da req-4a62f380-4bf9-43a7-9855-38bfb40c7934 service nova] Acquiring lock "refresh_cache-8f9aab3e-96ee-44c6-8d25-706a0ff857c4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.386741] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]520c3c63-994a-a775-0130-a2abaf117242, 'name': SearchDatastore_Task, 'duration_secs': 0.015555} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.386993] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.387306] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 1961c2c7-4400-4559-9f7c-bec8e8e263b0/1961c2c7-4400-4559-9f7c-bec8e8e263b0.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 835.387599] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d1941b19-e75c-45da-859a-9cf12c104f56 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.394648] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 835.394648] env[61987]: value = "task-1061750" [ 835.394648] env[61987]: _type = "Task" [ 835.394648] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.403567] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Acquiring lock "6ed3f0e9-3c39-4406-9863-931a117961a8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.403843] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061750, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.435637] env[61987]: DEBUG nova.network.neutron [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Updating instance_info_cache with network_info: [{"id": "ac11f37b-72cc-4fa6-a61a-4535486b9d7f", "address": "fa:16:3e:05:e0:5a", "network": {"id": "e2d0ce90-ff14-47de-bc4e-a2564d1268a7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-222691221-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d828262de7b14d2ab38193d7d34e8f7a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac11f37b-72", "ovs_interfaceid": "ac11f37b-72cc-4fa6-a61a-4535486b9d7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.462955] env[61987]: DEBUG nova.scheduler.client.report [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 835.548923] env[61987]: DEBUG oslo_vmware.api [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061747, 'name': PowerOnVM_Task, 'duration_secs': 0.862207} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.549252] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 835.549480] env[61987]: INFO nova.compute.manager [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Took 7.35 seconds to spawn the instance on the hypervisor. [ 835.549687] env[61987]: DEBUG nova.compute.manager [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 835.553528] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f2009c5-fa99-4261-8bcb-7f9e00b4f0d1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.591099] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c5b2768d-1a55-4e78-ab27-a92a45e7f2ab tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Lock "6ed3f0e9-3c39-4406-9863-931a117961a8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.460s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.591293] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Lock "6ed3f0e9-3c39-4406-9863-931a117961a8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.188s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.591544] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Acquiring lock "6ed3f0e9-3c39-4406-9863-931a117961a8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.591765] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Lock "6ed3f0e9-3c39-4406-9863-931a117961a8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.591956] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Lock "6ed3f0e9-3c39-4406-9863-931a117961a8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.595422] env[61987]: INFO nova.compute.manager [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Terminating instance [ 835.667800] env[61987]: DEBUG nova.network.neutron [-] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.907465] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061750, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.937941] env[61987]: DEBUG oslo_concurrency.lockutils [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Releasing lock "refresh_cache-8f9aab3e-96ee-44c6-8d25-706a0ff857c4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.938334] env[61987]: DEBUG nova.compute.manager [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Instance network_info: |[{"id": "ac11f37b-72cc-4fa6-a61a-4535486b9d7f", "address": "fa:16:3e:05:e0:5a", "network": {"id": "e2d0ce90-ff14-47de-bc4e-a2564d1268a7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-222691221-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d828262de7b14d2ab38193d7d34e8f7a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac11f37b-72", "ovs_interfaceid": "ac11f37b-72cc-4fa6-a61a-4535486b9d7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 835.938668] env[61987]: DEBUG oslo_concurrency.lockutils [req-2ba5f0d3-40f2-476d-8fe4-bbc840df40da req-4a62f380-4bf9-43a7-9855-38bfb40c7934 service nova] Acquired lock "refresh_cache-8f9aab3e-96ee-44c6-8d25-706a0ff857c4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.938861] env[61987]: DEBUG nova.network.neutron [req-2ba5f0d3-40f2-476d-8fe4-bbc840df40da req-4a62f380-4bf9-43a7-9855-38bfb40c7934 service nova] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Refreshing network info cache for port ac11f37b-72cc-4fa6-a61a-4535486b9d7f {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 835.944021] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:05:e0:5a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f52a458-d157-48a3-b4e2-b8cc0779afe2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ac11f37b-72cc-4fa6-a61a-4535486b9d7f', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 835.948750] env[61987]: DEBUG oslo.service.loopingcall [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 835.952018] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 835.952603] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-80704b3f-ab42-4c6d-b457-d07911dfb85e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.973912] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.367s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.974727] env[61987]: DEBUG nova.compute.manager [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 835.978986] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.664s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.978986] env[61987]: DEBUG nova.objects.instance [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lazy-loading 'resources' on Instance uuid 4fd8fc17-fb13-42e7-9879-9f120fdcb330 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 835.987883] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 835.987883] env[61987]: value = "task-1061751" [ 835.987883] env[61987]: _type = "Task" [ 835.987883] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.999799] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061751, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.078983] env[61987]: INFO nova.compute.manager [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Took 44.55 seconds to build instance. [ 836.100530] env[61987]: DEBUG nova.compute.manager [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 836.100872] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 836.102876] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f3c3ba0-6eba-46e8-827e-d9d87f5d107d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.111493] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 836.112344] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fdd10145-1c17-40de-af3a-aa913ba844bc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.118303] env[61987]: DEBUG oslo_vmware.api [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Waiting for the task: (returnval){ [ 836.118303] env[61987]: value = "task-1061752" [ 836.118303] env[61987]: _type = "Task" [ 836.118303] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.126382] env[61987]: DEBUG oslo_vmware.api [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': task-1061752, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.169151] env[61987]: INFO nova.compute.manager [-] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Took 1.02 seconds to deallocate network for instance. [ 836.210550] env[61987]: DEBUG nova.network.neutron [req-2ba5f0d3-40f2-476d-8fe4-bbc840df40da req-4a62f380-4bf9-43a7-9855-38bfb40c7934 service nova] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Updated VIF entry in instance network info cache for port ac11f37b-72cc-4fa6-a61a-4535486b9d7f. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 836.210931] env[61987]: DEBUG nova.network.neutron [req-2ba5f0d3-40f2-476d-8fe4-bbc840df40da req-4a62f380-4bf9-43a7-9855-38bfb40c7934 service nova] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Updating instance_info_cache with network_info: [{"id": "ac11f37b-72cc-4fa6-a61a-4535486b9d7f", "address": "fa:16:3e:05:e0:5a", "network": {"id": "e2d0ce90-ff14-47de-bc4e-a2564d1268a7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-222691221-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d828262de7b14d2ab38193d7d34e8f7a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac11f37b-72", "ovs_interfaceid": "ac11f37b-72cc-4fa6-a61a-4535486b9d7f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.371794] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 836.372597] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3da5df0a-6bbb-4fad-a502-dc4780a38891 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.384638] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Waiting for the task: (returnval){ [ 836.384638] env[61987]: value = "task-1061753" [ 836.384638] env[61987]: _type = "Task" [ 836.384638] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.126649] env[61987]: DEBUG nova.compute.utils [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 837.127430] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fca0dd5d-722c-4126-b12d-9c10ff925b24 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "056c181c-37da-49ac-ae1f-f7f2032c579a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.611s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.128184] env[61987]: DEBUG oslo_concurrency.lockutils [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.128563] env[61987]: DEBUG oslo_concurrency.lockutils [req-2ba5f0d3-40f2-476d-8fe4-bbc840df40da req-4a62f380-4bf9-43a7-9855-38bfb40c7934 service nova] Releasing lock "refresh_cache-8f9aab3e-96ee-44c6-8d25-706a0ff857c4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.129130] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.129337] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.129522] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.129704] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.129862] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.131021] env[61987]: DEBUG oslo_concurrency.lockutils [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "056c181c-37da-49ac-ae1f-f7f2032c579a" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.131222] env[61987]: DEBUG oslo_concurrency.lockutils [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "056c181c-37da-49ac-ae1f-f7f2032c579a" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.131392] env[61987]: INFO nova.compute.manager [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Shelving [ 837.137391] env[61987]: DEBUG nova.compute.manager [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 837.137391] env[61987]: DEBUG nova.network.neutron [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 837.146843] env[61987]: INFO nova.compute.manager [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Terminating instance [ 837.151571] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061753, 'name': PowerOffVM_Task, 'duration_secs': 0.353936} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.162597] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 837.163481] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061750, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.691649} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.163870] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061751, 'name': CreateVM_Task, 'duration_secs': 0.696097} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.164107] env[61987]: DEBUG oslo_vmware.api [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': task-1061752, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.164773] env[61987]: DEBUG nova.compute.manager [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 837.164959] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 837.167016] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-133fbd1e-10ad-49de-b614-26ecc09cffc8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.168315] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 1961c2c7-4400-4559-9f7c-bec8e8e263b0/1961c2c7-4400-4559-9f7c-bec8e8e263b0.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 837.168526] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 837.168693] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 837.169593] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cde7c573-d1b5-4d4a-bb9f-39a5892444bb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.172381] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a5232d9c-ff2f-4f30-a66f-fbf22b7bbca4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.174386] env[61987]: DEBUG oslo_concurrency.lockutils [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/92117f13-9c42-454c-87bb-8c7e90801e79" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.174548] env[61987]: DEBUG oslo_concurrency.lockutils [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquired lock "[datastore1] devstack-image-cache_base/92117f13-9c42-454c-87bb-8c7e90801e79" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.174914] env[61987]: DEBUG oslo_concurrency.lockutils [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/92117f13-9c42-454c-87bb-8c7e90801e79" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 837.175717] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49aae522-87a6-4368-98a5-105ead6060dd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.195564] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 837.199922] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a3b2169e-7535-471c-80ff-344ecd5912ca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.202363] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c98cc59-dd78-43a7-9acf-0beec07cb918 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.206041] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 837.206041] env[61987]: value = "task-1061754" [ 837.206041] env[61987]: _type = "Task" [ 837.206041] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.207681] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 837.207681] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52cc042e-ec3f-2bbc-2091-d13302ee61e5" [ 837.207681] env[61987]: _type = "Task" [ 837.207681] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.222706] env[61987]: DEBUG oslo_vmware.api [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 837.222706] env[61987]: value = "task-1061755" [ 837.222706] env[61987]: _type = "Task" [ 837.222706] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.224279] env[61987]: DEBUG nova.policy [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0945d6196b8c473da5308e6d784ab479', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b77fbbfc7f6437083f9f30e8fdeeb35', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 837.232445] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061754, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.236761] env[61987]: DEBUG oslo_concurrency.lockutils [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Releasing lock "[datastore1] devstack-image-cache_base/92117f13-9c42-454c-87bb-8c7e90801e79" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.237067] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Processing image 92117f13-9c42-454c-87bb-8c7e90801e79 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 837.237318] env[61987]: DEBUG oslo_concurrency.lockutils [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/92117f13-9c42-454c-87bb-8c7e90801e79/92117f13-9c42-454c-87bb-8c7e90801e79.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.237491] env[61987]: DEBUG oslo_concurrency.lockutils [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquired lock "[datastore1] devstack-image-cache_base/92117f13-9c42-454c-87bb-8c7e90801e79/92117f13-9c42-454c-87bb-8c7e90801e79.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.237720] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 837.243076] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1e286456-3a6c-4423-a0d6-33c83af4759e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.244913] env[61987]: DEBUG oslo_vmware.api [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061755, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.253661] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 837.254585] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-12ea51eb-a54a-430e-8b90-abf4e8dd9e7a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.256217] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 837.256469] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 837.259430] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80d9b876-4cec-451c-8630-1560c47aa63f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.265567] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 837.265567] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52fdc9d1-5156-bd61-1d0b-2fc98ce17f29" [ 837.265567] env[61987]: _type = "Task" [ 837.265567] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.267220] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Waiting for the task: (returnval){ [ 837.267220] env[61987]: value = "task-1061756" [ 837.267220] env[61987]: _type = "Task" [ 837.267220] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.284590] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Preparing fetch location {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 837.284845] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Fetch image to [datastore1] OSTACK_IMG_1d86b010-a08b-40ca-baa2-3164788f5912/OSTACK_IMG_1d86b010-a08b-40ca-baa2-3164788f5912.vmdk {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 837.285611] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Downloading stream optimized image 92117f13-9c42-454c-87bb-8c7e90801e79 to [datastore1] OSTACK_IMG_1d86b010-a08b-40ca-baa2-3164788f5912/OSTACK_IMG_1d86b010-a08b-40ca-baa2-3164788f5912.vmdk on the data store datastore1 as vApp {{(pid=61987) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 837.285611] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Downloading image file data 92117f13-9c42-454c-87bb-8c7e90801e79 to the ESX as VM named 'OSTACK_IMG_1d86b010-a08b-40ca-baa2-3164788f5912' {{(pid=61987) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 837.288727] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] VM already powered off {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 837.288727] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 837.288727] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.288727] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.288727] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 837.288727] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1e7a728d-bb96-41cd-8638-2feb9196ffc8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.329548] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 837.330823] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 837.330823] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ec9ff4a-6642-44bc-b0d7-2cfb2813cfd4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.336506] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Waiting for the task: (returnval){ [ 837.336506] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]523fd983-46ee-3086-6b9d-c4883f0cca63" [ 837.336506] env[61987]: _type = "Task" [ 837.336506] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.356383] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523fd983-46ee-3086-6b9d-c4883f0cca63, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.385900] env[61987]: DEBUG oslo_vmware.rw_handles [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 837.385900] env[61987]: value = "resgroup-9" [ 837.385900] env[61987]: _type = "ResourcePool" [ 837.385900] env[61987]: }. {{(pid=61987) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 837.386263] env[61987]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-a8821bc4-ae36-4f2b-8779-39646af03009 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.411559] env[61987]: DEBUG oslo_vmware.rw_handles [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lease: (returnval){ [ 837.411559] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52aed0e8-b413-1876-1d2f-6df92a2bdbda" [ 837.411559] env[61987]: _type = "HttpNfcLease" [ 837.411559] env[61987]: } obtained for vApp import into resource pool (val){ [ 837.411559] env[61987]: value = "resgroup-9" [ 837.411559] env[61987]: _type = "ResourcePool" [ 837.411559] env[61987]: }. {{(pid=61987) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 837.411839] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the lease: (returnval){ [ 837.411839] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52aed0e8-b413-1876-1d2f-6df92a2bdbda" [ 837.411839] env[61987]: _type = "HttpNfcLease" [ 837.411839] env[61987]: } to be ready. {{(pid=61987) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 837.419636] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 837.419636] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52aed0e8-b413-1876-1d2f-6df92a2bdbda" [ 837.419636] env[61987]: _type = "HttpNfcLease" [ 837.419636] env[61987]: } is initializing. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 837.494039] env[61987]: DEBUG oslo_concurrency.lockutils [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.494323] env[61987]: DEBUG oslo_concurrency.lockutils [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.497050] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40e1f308-887f-43cd-977e-0904fa58f338 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.504526] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb72a00f-89d5-47c2-8887-85eb4cfd3174 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.536042] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad6ae37-8605-4d30-b16a-be5865c2cb58 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.544057] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5fc4c98-0618-4b84-92a1-e0a0d40aedb7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.561846] env[61987]: DEBUG nova.compute.provider_tree [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.634832] env[61987]: DEBUG nova.compute.manager [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 837.640027] env[61987]: DEBUG oslo_vmware.api [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': task-1061752, 'name': PowerOffVM_Task, 'duration_secs': 1.347169} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.640119] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 837.640301] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 837.640559] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ab04086c-aa99-4967-b824-60a327d7fe07 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.655404] env[61987]: DEBUG nova.network.neutron [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Successfully created port: 0a4d6904-241e-451e-86d4-2cec32cf3b25 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 837.717662] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 837.718022] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 837.718087] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Deleting the datastore file [datastore2] 6ed3f0e9-3c39-4406-9863-931a117961a8 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 837.722531] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-307d2011-b59c-4521-9bfd-b8f3de997d26 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.724300] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061754, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079554} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.724564] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 837.725758] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec5c1087-daee-4bc2-8291-ca4da7cfb747 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.729206] env[61987]: DEBUG oslo_vmware.api [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Waiting for the task: (returnval){ [ 837.729206] env[61987]: value = "task-1061759" [ 837.729206] env[61987]: _type = "Task" [ 837.729206] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.753495] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 1961c2c7-4400-4559-9f7c-bec8e8e263b0/1961c2c7-4400-4559-9f7c-bec8e8e263b0.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 837.757380] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0adecc6e-5b5a-41e9-a525-e8ada8b5bb1c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.773789] env[61987]: DEBUG oslo_vmware.api [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061755, 'name': PowerOffVM_Task, 'duration_secs': 0.19825} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.778380] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 837.778380] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 837.778380] env[61987]: DEBUG oslo_vmware.api [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': task-1061759, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.778588] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7c13ddca-8ac7-478c-9c0c-b681ea611adc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.783476] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 837.783476] env[61987]: value = "task-1061760" [ 837.783476] env[61987]: _type = "Task" [ 837.783476] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.791611] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061760, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.842183] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 837.843525] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 837.843525] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Deleting the datastore file [datastore2] 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 837.843525] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eb68b2cd-e3e7-4f39-8e26-464e42a746ed {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.848816] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523fd983-46ee-3086-6b9d-c4883f0cca63, 'name': SearchDatastore_Task, 'duration_secs': 0.020864} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.850598] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f19b4158-c9e9-4d70-8c6a-098cf5b97fb9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.853993] env[61987]: DEBUG oslo_vmware.api [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 837.853993] env[61987]: value = "task-1061762" [ 837.853993] env[61987]: _type = "Task" [ 837.853993] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.858484] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Waiting for the task: (returnval){ [ 837.858484] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52231143-0d86-708f-b893-2caafe1cb749" [ 837.858484] env[61987]: _type = "Task" [ 837.858484] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.864951] env[61987]: DEBUG oslo_vmware.api [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061762, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.870252] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52231143-0d86-708f-b893-2caafe1cb749, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.920616] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 837.920616] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52aed0e8-b413-1876-1d2f-6df92a2bdbda" [ 837.920616] env[61987]: _type = "HttpNfcLease" [ 837.920616] env[61987]: } is initializing. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 838.000824] env[61987]: DEBUG nova.compute.manager [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 838.064936] env[61987]: DEBUG nova.scheduler.client.report [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 838.153840] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 838.154208] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8adc29b6-e716-4446-936e-c5b42bc1fe5e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.161688] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 838.161688] env[61987]: value = "task-1061763" [ 838.161688] env[61987]: _type = "Task" [ 838.161688] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.171811] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061763, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.243580] env[61987]: DEBUG oslo_vmware.api [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Task: {'id': task-1061759, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168186} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.243889] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 838.244105] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 838.244333] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 838.244537] env[61987]: INFO nova.compute.manager [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Took 2.14 seconds to destroy the instance on the hypervisor. [ 838.244796] env[61987]: DEBUG oslo.service.loopingcall [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 838.244996] env[61987]: DEBUG nova.compute.manager [-] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 838.245107] env[61987]: DEBUG nova.network.neutron [-] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 838.296723] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061760, 'name': ReconfigVM_Task, 'duration_secs': 0.320212} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.297305] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 1961c2c7-4400-4559-9f7c-bec8e8e263b0/1961c2c7-4400-4559-9f7c-bec8e8e263b0.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 838.301399] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8d9bdbd1-6ff7-4a65-b612-a2393e6d1fd4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.309202] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 838.309202] env[61987]: value = "task-1061764" [ 838.309202] env[61987]: _type = "Task" [ 838.309202] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.320329] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061764, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.367586] env[61987]: DEBUG oslo_vmware.api [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061762, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176428} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.368337] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 838.368631] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 838.368739] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 838.368934] env[61987]: INFO nova.compute.manager [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Took 1.20 seconds to destroy the instance on the hypervisor. [ 838.369311] env[61987]: DEBUG oslo.service.loopingcall [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 838.374947] env[61987]: DEBUG nova.compute.manager [-] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 838.375282] env[61987]: DEBUG nova.network.neutron [-] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 838.380172] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52231143-0d86-708f-b893-2caafe1cb749, 'name': SearchDatastore_Task, 'duration_secs': 0.021987} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.380510] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.380800] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 98d895ee-1d1f-49cd-a055-fe6c8462a692/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk. {{(pid=61987) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 838.381506] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cf336684-e3cf-4522-862d-fc8a6aba63f4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.391150] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Waiting for the task: (returnval){ [ 838.391150] env[61987]: value = "task-1061765" [ 838.391150] env[61987]: _type = "Task" [ 838.391150] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.401015] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061765, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.423586] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 838.423586] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52aed0e8-b413-1876-1d2f-6df92a2bdbda" [ 838.423586] env[61987]: _type = "HttpNfcLease" [ 838.423586] env[61987]: } is ready. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 838.423870] env[61987]: DEBUG oslo_vmware.rw_handles [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 838.423870] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52aed0e8-b413-1876-1d2f-6df92a2bdbda" [ 838.423870] env[61987]: _type = "HttpNfcLease" [ 838.423870] env[61987]: }. {{(pid=61987) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 838.424806] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca7e94b1-bbf1-4e67-b979-13cdee6aeca6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.435067] env[61987]: DEBUG oslo_vmware.rw_handles [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb6981-5892-9359-3298-17ee5dea2706/disk-0.vmdk from lease info. {{(pid=61987) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 838.435067] env[61987]: DEBUG oslo_vmware.rw_handles [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb6981-5892-9359-3298-17ee5dea2706/disk-0.vmdk. {{(pid=61987) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 838.506988] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-347a5588-4a34-4aa5-b822-51b137335968 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.535287] env[61987]: DEBUG oslo_concurrency.lockutils [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.571470] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.593s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.576174] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.105s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.578157] env[61987]: DEBUG nova.objects.instance [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lazy-loading 'resources' on Instance uuid 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 838.604829] env[61987]: INFO nova.scheduler.client.report [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Deleted allocations for instance 4fd8fc17-fb13-42e7-9879-9f120fdcb330 [ 838.652189] env[61987]: DEBUG nova.compute.manager [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 838.673981] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061763, 'name': PowerOffVM_Task, 'duration_secs': 0.175029} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.674316] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 838.675513] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f211368-b61a-4def-b6ae-3699e52e33e3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.700914] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75288b27-8ebc-4a7c-bb81-497d1c5855c6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.712610] env[61987]: DEBUG nova.virt.hardware [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 838.712894] env[61987]: DEBUG nova.virt.hardware [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 838.713071] env[61987]: DEBUG nova.virt.hardware [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 838.713283] env[61987]: DEBUG nova.virt.hardware [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 838.713435] env[61987]: DEBUG nova.virt.hardware [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 838.713588] env[61987]: DEBUG nova.virt.hardware [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 838.713797] env[61987]: DEBUG nova.virt.hardware [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 838.713959] env[61987]: DEBUG nova.virt.hardware [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 838.714258] env[61987]: DEBUG nova.virt.hardware [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 838.714481] env[61987]: DEBUG nova.virt.hardware [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 838.714676] env[61987]: DEBUG nova.virt.hardware [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 838.717936] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9b07109-95e5-487f-94d9-ed8bd67facde {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.732735] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c757efe-d212-4213-88a4-6131cce7a4fc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.788401] env[61987]: DEBUG nova.compute.manager [req-92fe78b9-f2b7-4490-a3c8-633aeb408298 req-3312af1c-92c5-4079-b574-aeef4ab062fc service nova] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Received event network-vif-deleted-6119ef11-a392-4c05-8c71-6e5b25f3cdc9 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 838.788642] env[61987]: INFO nova.compute.manager [req-92fe78b9-f2b7-4490-a3c8-633aeb408298 req-3312af1c-92c5-4079-b574-aeef4ab062fc service nova] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Neutron deleted interface 6119ef11-a392-4c05-8c71-6e5b25f3cdc9; detaching it from the instance and deleting it from the info cache [ 838.788830] env[61987]: DEBUG nova.network.neutron [req-92fe78b9-f2b7-4490-a3c8-633aeb408298 req-3312af1c-92c5-4079-b574-aeef4ab062fc service nova] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.821951] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061764, 'name': Rename_Task, 'duration_secs': 0.155653} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.822319] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 838.823815] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-af5a687f-e084-4fd0-bc31-ed0cb764b412 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.833058] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 838.833058] env[61987]: value = "task-1061766" [ 838.833058] env[61987]: _type = "Task" [ 838.833058] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.842918] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061766, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.904237] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061765, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.944725] env[61987]: DEBUG nova.compute.manager [req-b9488d59-e5b6-4254-a75e-c21c1eafa9dd req-47791ff0-07cf-42bd-bc97-f4f2ce758a79 service nova] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Received event network-vif-deleted-55e46e63-9076-400e-a73a-19fc4bb21bec {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 838.945017] env[61987]: INFO nova.compute.manager [req-b9488d59-e5b6-4254-a75e-c21c1eafa9dd req-47791ff0-07cf-42bd-bc97-f4f2ce758a79 service nova] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Neutron deleted interface 55e46e63-9076-400e-a73a-19fc4bb21bec; detaching it from the instance and deleting it from the info cache [ 838.945634] env[61987]: DEBUG nova.network.neutron [req-b9488d59-e5b6-4254-a75e-c21c1eafa9dd req-47791ff0-07cf-42bd-bc97-f4f2ce758a79 service nova] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.001908] env[61987]: DEBUG oslo_concurrency.lockutils [None req-23ba0751-06ef-4b8e-8cb8-fce4bf326292 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "ad40937f-d5a9-4708-8b81-06087d38d765" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.002323] env[61987]: DEBUG oslo_concurrency.lockutils [None req-23ba0751-06ef-4b8e-8cb8-fce4bf326292 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ad40937f-d5a9-4708-8b81-06087d38d765" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.089484] env[61987]: DEBUG nova.network.neutron [-] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.118308] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b6a3753-5ab5-4fbf-a079-e7f509942886 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "4fd8fc17-fb13-42e7-9879-9f120fdcb330" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.170s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.224996] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Creating Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 839.227139] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a05541b9-822d-48bf-9b44-342af1bf2d0d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.229094] env[61987]: DEBUG nova.network.neutron [-] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.235478] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 839.235478] env[61987]: value = "task-1061767" [ 839.235478] env[61987]: _type = "Task" [ 839.235478] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.247970] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061767, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.293347] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7922e4d9-0811-48c9-b9e9-dc3235cc60a4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.307944] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e80d5469-16f2-495d-9e75-96491fc85b51 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.357729] env[61987]: DEBUG nova.compute.manager [req-92fe78b9-f2b7-4490-a3c8-633aeb408298 req-3312af1c-92c5-4079-b574-aeef4ab062fc service nova] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Detach interface failed, port_id=6119ef11-a392-4c05-8c71-6e5b25f3cdc9, reason: Instance 6ed3f0e9-3c39-4406-9863-931a117961a8 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 839.358278] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061766, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.361944] env[61987]: DEBUG oslo_vmware.rw_handles [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Completed reading data from the image iterator. {{(pid=61987) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 839.362208] env[61987]: DEBUG oslo_vmware.rw_handles [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb6981-5892-9359-3298-17ee5dea2706/disk-0.vmdk. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 839.363132] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69e93612-5257-46ca-9c37-8036d90341af {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.370110] env[61987]: DEBUG oslo_vmware.rw_handles [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb6981-5892-9359-3298-17ee5dea2706/disk-0.vmdk is in state: ready. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 839.370712] env[61987]: DEBUG oslo_vmware.rw_handles [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb6981-5892-9359-3298-17ee5dea2706/disk-0.vmdk. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 839.370712] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-220e3b74-7eb2-43ad-b956-623d4f60abae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.405013] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061765, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.606583} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.405401] env[61987]: INFO nova.virt.vmwareapi.ds_util [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 98d895ee-1d1f-49cd-a055-fe6c8462a692/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk. [ 839.406301] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8539433-b911-4aec-8e4f-ec881d69113c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.437487] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 98d895ee-1d1f-49cd-a055-fe6c8462a692/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 839.441614] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-19955fdb-da02-4f92-bc0f-39b7d6b35e88 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.459561] env[61987]: DEBUG nova.network.neutron [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Successfully updated port: 0a4d6904-241e-451e-86d4-2cec32cf3b25 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 839.461498] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5710c62a-3d57-4987-87c4-6360980e5d4d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.473411] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a00bb6-fee8-41d2-82a6-026cc1a5d119 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.484334] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Waiting for the task: (returnval){ [ 839.484334] env[61987]: value = "task-1061768" [ 839.484334] env[61987]: _type = "Task" [ 839.484334] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.487253] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8fd3248-b663-49b3-9a81-4c17f22999b3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.501428] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061768, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.512218] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-796ce77f-5955-4545-b616-30a09595c64c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.516016] env[61987]: INFO nova.compute.manager [None req-23ba0751-06ef-4b8e-8cb8-fce4bf326292 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Detaching volume 26cf07f3-1a6a-47ba-bd88-c91cccac13bc [ 839.519257] env[61987]: DEBUG nova.compute.manager [req-b9488d59-e5b6-4254-a75e-c21c1eafa9dd req-47791ff0-07cf-42bd-bc97-f4f2ce758a79 service nova] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Detach interface failed, port_id=55e46e63-9076-400e-a73a-19fc4bb21bec, reason: Instance 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 839.549992] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fdd181e-3df6-48c9-9b10-6f08b40c60f0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.557578] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68a6c70c-3c56-46e3-8d1e-0fe07d9f12bd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.571437] env[61987]: DEBUG nova.compute.provider_tree [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.590475] env[61987]: INFO nova.virt.block_device [None req-23ba0751-06ef-4b8e-8cb8-fce4bf326292 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Attempting to driver detach volume 26cf07f3-1a6a-47ba-bd88-c91cccac13bc from mountpoint /dev/sdb [ 839.590768] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-23ba0751-06ef-4b8e-8cb8-fce4bf326292 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Volume detach. Driver type: vmdk {{(pid=61987) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 839.591026] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-23ba0751-06ef-4b8e-8cb8-fce4bf326292 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234324', 'volume_id': '26cf07f3-1a6a-47ba-bd88-c91cccac13bc', 'name': 'volume-26cf07f3-1a6a-47ba-bd88-c91cccac13bc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ad40937f-d5a9-4708-8b81-06087d38d765', 'attached_at': '', 'detached_at': '', 'volume_id': '26cf07f3-1a6a-47ba-bd88-c91cccac13bc', 'serial': '26cf07f3-1a6a-47ba-bd88-c91cccac13bc'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 839.591468] env[61987]: INFO nova.compute.manager [-] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Took 1.35 seconds to deallocate network for instance. [ 839.592362] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeecd604-c95b-446b-a739-a08c1f65d1fe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.619424] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c6f104-9401-4493-b1be-0f1de8362a69 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.627196] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8f39d16-1588-467d-8002-17822d1aa980 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.649024] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f31c217-bc67-47c7-9012-5a558296143a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.665583] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-23ba0751-06ef-4b8e-8cb8-fce4bf326292 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] The volume has not been displaced from its original location: [datastore1] volume-26cf07f3-1a6a-47ba-bd88-c91cccac13bc/volume-26cf07f3-1a6a-47ba-bd88-c91cccac13bc.vmdk. No consolidation needed. {{(pid=61987) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 839.670825] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-23ba0751-06ef-4b8e-8cb8-fce4bf326292 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Reconfiguring VM instance instance-00000031 to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 839.671590] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dfa6aa2d-b5fb-4e8e-8019-84f1a43ccf7c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.687899] env[61987]: DEBUG oslo_vmware.rw_handles [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52eb6981-5892-9359-3298-17ee5dea2706/disk-0.vmdk. {{(pid=61987) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 839.688609] env[61987]: INFO nova.virt.vmwareapi.images [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Downloaded image file data 92117f13-9c42-454c-87bb-8c7e90801e79 [ 839.690037] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b6d066a-a003-48df-bd79-151c65af6325 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.692923] env[61987]: DEBUG oslo_vmware.api [None req-23ba0751-06ef-4b8e-8cb8-fce4bf326292 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 839.692923] env[61987]: value = "task-1061769" [ 839.692923] env[61987]: _type = "Task" [ 839.692923] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.707208] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-014a741a-3b0e-4d00-b6c2-1e948bd030c5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.712046] env[61987]: DEBUG oslo_vmware.api [None req-23ba0751-06ef-4b8e-8cb8-fce4bf326292 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061769, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.735260] env[61987]: INFO nova.compute.manager [-] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Took 1.36 seconds to deallocate network for instance. [ 839.749796] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061767, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.828200] env[61987]: INFO nova.virt.vmwareapi.images [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] The imported VM was unregistered [ 839.830485] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Caching image {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 839.830850] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Creating directory with path [datastore1] devstack-image-cache_base/92117f13-9c42-454c-87bb-8c7e90801e79 {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 839.831185] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-de4b4601-803c-4ee5-8284-a326243aa524 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.845270] env[61987]: DEBUG oslo_vmware.api [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061766, 'name': PowerOnVM_Task, 'duration_secs': 0.593436} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.846398] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 839.846791] env[61987]: INFO nova.compute.manager [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Took 9.13 seconds to spawn the instance on the hypervisor. [ 839.846983] env[61987]: DEBUG nova.compute.manager [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 839.847292] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Created directory with path [datastore1] devstack-image-cache_base/92117f13-9c42-454c-87bb-8c7e90801e79 {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 839.847470] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_1d86b010-a08b-40ca-baa2-3164788f5912/OSTACK_IMG_1d86b010-a08b-40ca-baa2-3164788f5912.vmdk to [datastore1] devstack-image-cache_base/92117f13-9c42-454c-87bb-8c7e90801e79/92117f13-9c42-454c-87bb-8c7e90801e79.vmdk. {{(pid=61987) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 839.848206] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23257af7-5745-4596-a0bc-06c53d24abfc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.850630] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-18bc7b05-4a02-4a8c-a87e-f8532307520f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.861973] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 839.861973] env[61987]: value = "task-1061771" [ 839.861973] env[61987]: _type = "Task" [ 839.861973] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.869896] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061771, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.962658] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "refresh_cache-742297eb-c45d-4f45-85d2-e9bdd3106d1e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.962819] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquired lock "refresh_cache-742297eb-c45d-4f45-85d2-e9bdd3106d1e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.962977] env[61987]: DEBUG nova.network.neutron [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 839.996253] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061768, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.074395] env[61987]: DEBUG nova.scheduler.client.report [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 840.121065] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.204584] env[61987]: DEBUG oslo_vmware.api [None req-23ba0751-06ef-4b8e-8cb8-fce4bf326292 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061769, 'name': ReconfigVM_Task, 'duration_secs': 0.269293} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.205203] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-23ba0751-06ef-4b8e-8cb8-fce4bf326292 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Reconfigured VM instance instance-00000031 to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 840.210443] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d55f3326-3361-4bb4-9c36-58ce4255727b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.228056] env[61987]: DEBUG oslo_vmware.api [None req-23ba0751-06ef-4b8e-8cb8-fce4bf326292 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 840.228056] env[61987]: value = "task-1061772" [ 840.228056] env[61987]: _type = "Task" [ 840.228056] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.236565] env[61987]: DEBUG oslo_vmware.api [None req-23ba0751-06ef-4b8e-8cb8-fce4bf326292 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061772, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.246493] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.246812] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061767, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.372852] env[61987]: INFO nova.compute.manager [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Took 38.55 seconds to build instance. [ 840.377039] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061771, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.499168] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061768, 'name': ReconfigVM_Task, 'duration_secs': 0.821704} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.499683] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 98d895ee-1d1f-49cd-a055-fe6c8462a692/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 840.501637] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8260117-5a21-4de8-8eb5-c996199247ec {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.534085] env[61987]: DEBUG nova.network.neutron [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 840.537762] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29f897a8-91b3-4716-8241-2db09b9ba455 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.552811] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "1bba4036-5211-4a43-b2e6-5276851335e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.553123] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "1bba4036-5211-4a43-b2e6-5276851335e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.560901] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Waiting for the task: (returnval){ [ 840.560901] env[61987]: value = "task-1061773" [ 840.560901] env[61987]: _type = "Task" [ 840.560901] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.572253] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061773, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.580420] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.004s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.587037] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 9.161s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.587406] env[61987]: DEBUG nova.objects.instance [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61987) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 840.626982] env[61987]: INFO nova.scheduler.client.report [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Deleted allocations for instance 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6 [ 840.740844] env[61987]: DEBUG oslo_vmware.api [None req-23ba0751-06ef-4b8e-8cb8-fce4bf326292 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061772, 'name': ReconfigVM_Task, 'duration_secs': 0.149608} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.744763] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-23ba0751-06ef-4b8e-8cb8-fce4bf326292 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234324', 'volume_id': '26cf07f3-1a6a-47ba-bd88-c91cccac13bc', 'name': 'volume-26cf07f3-1a6a-47ba-bd88-c91cccac13bc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ad40937f-d5a9-4708-8b81-06087d38d765', 'attached_at': '', 'detached_at': '', 'volume_id': '26cf07f3-1a6a-47ba-bd88-c91cccac13bc', 'serial': '26cf07f3-1a6a-47ba-bd88-c91cccac13bc'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 840.753890] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061767, 'name': CreateSnapshot_Task, 'duration_secs': 1.015789} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.754437] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Created Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 840.755464] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd030d02-fd94-49cc-99fa-12148343efdd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.828438] env[61987]: DEBUG nova.network.neutron [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Updating instance_info_cache with network_info: [{"id": "0a4d6904-241e-451e-86d4-2cec32cf3b25", "address": "fa:16:3e:bb:78:bb", "network": {"id": "26d453d1-8edf-45e1-9e49-2a3225de6633", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1400564876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b77fbbfc7f6437083f9f30e8fdeeb35", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a4d6904-24", "ovs_interfaceid": "0a4d6904-241e-451e-86d4-2cec32cf3b25", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.877830] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6700c60d-a4fb-4c5c-837e-230c3128441c tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "1961c2c7-4400-4559-9f7c-bec8e8e263b0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.071s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.878557] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061771, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.952401] env[61987]: DEBUG nova.compute.manager [req-3afec384-6dae-4d41-a08c-fca3bfa4562b req-d94d99d5-c486-4b83-915e-ad95d00efc05 service nova] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Received event network-vif-plugged-0a4d6904-241e-451e-86d4-2cec32cf3b25 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 840.952639] env[61987]: DEBUG oslo_concurrency.lockutils [req-3afec384-6dae-4d41-a08c-fca3bfa4562b req-d94d99d5-c486-4b83-915e-ad95d00efc05 service nova] Acquiring lock "742297eb-c45d-4f45-85d2-e9bdd3106d1e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.952861] env[61987]: DEBUG oslo_concurrency.lockutils [req-3afec384-6dae-4d41-a08c-fca3bfa4562b req-d94d99d5-c486-4b83-915e-ad95d00efc05 service nova] Lock "742297eb-c45d-4f45-85d2-e9bdd3106d1e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.953122] env[61987]: DEBUG oslo_concurrency.lockutils [req-3afec384-6dae-4d41-a08c-fca3bfa4562b req-d94d99d5-c486-4b83-915e-ad95d00efc05 service nova] Lock "742297eb-c45d-4f45-85d2-e9bdd3106d1e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.953323] env[61987]: DEBUG nova.compute.manager [req-3afec384-6dae-4d41-a08c-fca3bfa4562b req-d94d99d5-c486-4b83-915e-ad95d00efc05 service nova] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] No waiting events found dispatching network-vif-plugged-0a4d6904-241e-451e-86d4-2cec32cf3b25 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 840.953512] env[61987]: WARNING nova.compute.manager [req-3afec384-6dae-4d41-a08c-fca3bfa4562b req-d94d99d5-c486-4b83-915e-ad95d00efc05 service nova] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Received unexpected event network-vif-plugged-0a4d6904-241e-451e-86d4-2cec32cf3b25 for instance with vm_state building and task_state spawning. [ 840.953668] env[61987]: DEBUG nova.compute.manager [req-3afec384-6dae-4d41-a08c-fca3bfa4562b req-d94d99d5-c486-4b83-915e-ad95d00efc05 service nova] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Received event network-changed-0a4d6904-241e-451e-86d4-2cec32cf3b25 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 840.953828] env[61987]: DEBUG nova.compute.manager [req-3afec384-6dae-4d41-a08c-fca3bfa4562b req-d94d99d5-c486-4b83-915e-ad95d00efc05 service nova] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Refreshing instance network info cache due to event network-changed-0a4d6904-241e-451e-86d4-2cec32cf3b25. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 840.954009] env[61987]: DEBUG oslo_concurrency.lockutils [req-3afec384-6dae-4d41-a08c-fca3bfa4562b req-d94d99d5-c486-4b83-915e-ad95d00efc05 service nova] Acquiring lock "refresh_cache-742297eb-c45d-4f45-85d2-e9bdd3106d1e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.056348] env[61987]: DEBUG nova.compute.manager [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 841.071231] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061773, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.140994] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a8520b37-36db-4465-ae18-2ff3cb96440a tempest-MigrationsAdminTest-2073369025 tempest-MigrationsAdminTest-2073369025-project-member] Lock "80f13fca-2df1-4aa0-96f1-b61a60c4fdc6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.040s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.277325] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Creating linked-clone VM from snapshot {{(pid=61987) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 841.277720] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d1de70ac-6bb3-4297-94e9-fbe99edb91ab {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.288863] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 841.288863] env[61987]: value = "task-1061774" [ 841.288863] env[61987]: _type = "Task" [ 841.288863] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.302600] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061774, 'name': CloneVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.306696] env[61987]: DEBUG nova.objects.instance [None req-23ba0751-06ef-4b8e-8cb8-fce4bf326292 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lazy-loading 'flavor' on Instance uuid ad40937f-d5a9-4708-8b81-06087d38d765 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 841.333392] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Releasing lock "refresh_cache-742297eb-c45d-4f45-85d2-e9bdd3106d1e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.333392] env[61987]: DEBUG nova.compute.manager [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Instance network_info: |[{"id": "0a4d6904-241e-451e-86d4-2cec32cf3b25", "address": "fa:16:3e:bb:78:bb", "network": {"id": "26d453d1-8edf-45e1-9e49-2a3225de6633", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1400564876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b77fbbfc7f6437083f9f30e8fdeeb35", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a4d6904-24", "ovs_interfaceid": "0a4d6904-241e-451e-86d4-2cec32cf3b25", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 841.333392] env[61987]: DEBUG oslo_concurrency.lockutils [req-3afec384-6dae-4d41-a08c-fca3bfa4562b req-d94d99d5-c486-4b83-915e-ad95d00efc05 service nova] Acquired lock "refresh_cache-742297eb-c45d-4f45-85d2-e9bdd3106d1e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.333392] env[61987]: DEBUG nova.network.neutron [req-3afec384-6dae-4d41-a08c-fca3bfa4562b req-d94d99d5-c486-4b83-915e-ad95d00efc05 service nova] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Refreshing network info cache for port 0a4d6904-241e-451e-86d4-2cec32cf3b25 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 841.333392] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bb:78:bb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3734b156-0f7d-4721-b23c-d000412ec2eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0a4d6904-241e-451e-86d4-2cec32cf3b25', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 841.342847] env[61987]: DEBUG oslo.service.loopingcall [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.343899] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 841.344272] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a33fc9c8-d7f2-43cd-8848-dc7a1c0c2d8c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.364754] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 841.364754] env[61987]: value = "task-1061775" [ 841.364754] env[61987]: _type = "Task" [ 841.364754] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.376536] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061775, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.380286] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061771, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.525534] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "1961c2c7-4400-4559-9f7c-bec8e8e263b0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.525821] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "1961c2c7-4400-4559-9f7c-bec8e8e263b0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.526118] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "1961c2c7-4400-4559-9f7c-bec8e8e263b0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.526889] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "1961c2c7-4400-4559-9f7c-bec8e8e263b0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.526889] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "1961c2c7-4400-4559-9f7c-bec8e8e263b0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.531696] env[61987]: INFO nova.compute.manager [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Terminating instance [ 841.579771] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061773, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.584267] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.601024] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7109e882-e51c-4b41-8870-94ed6394b1ce tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.601024] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.243s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.601024] env[61987]: DEBUG nova.objects.instance [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lazy-loading 'resources' on Instance uuid 4edd0159-6db1-41bd-a616-31a34e984059 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 841.804312] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061774, 'name': CloneVM_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.880530] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061771, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.885044] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061775, 'name': CreateVM_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.038874] env[61987]: DEBUG nova.compute.manager [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 842.039178] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 842.040687] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-738a7f6f-75c1-4e6f-aefd-8e6d4a3d73c4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.055336] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 842.056337] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-282bff8a-f588-43ea-9582-716bf3cea8e8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.070247] env[61987]: DEBUG oslo_vmware.api [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 842.070247] env[61987]: value = "task-1061776" [ 842.070247] env[61987]: _type = "Task" [ 842.070247] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.084766] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061773, 'name': ReconfigVM_Task, 'duration_secs': 1.247311} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.085895] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 842.086298] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2af6913d-026f-4b83-b07a-36497d472998 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.093987] env[61987]: DEBUG oslo_vmware.api [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061776, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.099531] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Waiting for the task: (returnval){ [ 842.099531] env[61987]: value = "task-1061777" [ 842.099531] env[61987]: _type = "Task" [ 842.099531] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.116033] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061777, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.166727] env[61987]: DEBUG nova.network.neutron [req-3afec384-6dae-4d41-a08c-fca3bfa4562b req-d94d99d5-c486-4b83-915e-ad95d00efc05 service nova] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Updated VIF entry in instance network info cache for port 0a4d6904-241e-451e-86d4-2cec32cf3b25. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 842.167494] env[61987]: DEBUG nova.network.neutron [req-3afec384-6dae-4d41-a08c-fca3bfa4562b req-d94d99d5-c486-4b83-915e-ad95d00efc05 service nova] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Updating instance_info_cache with network_info: [{"id": "0a4d6904-241e-451e-86d4-2cec32cf3b25", "address": "fa:16:3e:bb:78:bb", "network": {"id": "26d453d1-8edf-45e1-9e49-2a3225de6633", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1400564876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b77fbbfc7f6437083f9f30e8fdeeb35", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a4d6904-24", "ovs_interfaceid": "0a4d6904-241e-451e-86d4-2cec32cf3b25", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.303514] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061774, 'name': CloneVM_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.312629] env[61987]: DEBUG oslo_concurrency.lockutils [None req-23ba0751-06ef-4b8e-8cb8-fce4bf326292 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ad40937f-d5a9-4708-8b81-06087d38d765" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.310s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.383923] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061775, 'name': CreateVM_Task, 'duration_secs': 0.750874} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.386907] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 842.387261] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061771, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.390538] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.390732] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.391105] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 842.392042] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d0cd4bc-faba-4e73-8fa1-d8eccba8927f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.398972] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 842.398972] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5206f881-fc99-21bc-36a5-5fe15f0afaf8" [ 842.398972] env[61987]: _type = "Task" [ 842.398972] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.414360] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5206f881-fc99-21bc-36a5-5fe15f0afaf8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.416641] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e858bbc7-8e28-408a-8fec-44d58abffa92 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.422093] env[61987]: DEBUG oslo_concurrency.lockutils [None req-73538a34-1b2f-4a0d-b206-72fe3e124000 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "ad40937f-d5a9-4708-8b81-06087d38d765" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.422365] env[61987]: DEBUG oslo_concurrency.lockutils [None req-73538a34-1b2f-4a0d-b206-72fe3e124000 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ad40937f-d5a9-4708-8b81-06087d38d765" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.422566] env[61987]: DEBUG nova.compute.manager [None req-73538a34-1b2f-4a0d-b206-72fe3e124000 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 842.425433] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5249728d-40a1-4c48-8712-e2ee649de966 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.429395] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c610c5d-c1d4-4f89-8519-38a38148022c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.433314] env[61987]: DEBUG oslo_concurrency.lockutils [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "51a36102-795e-47b7-b96a-857e54dc703e" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.433498] env[61987]: DEBUG oslo_concurrency.lockutils [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "51a36102-795e-47b7-b96a-857e54dc703e" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.433685] env[61987]: INFO nova.compute.manager [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Shelving [ 842.440391] env[61987]: DEBUG nova.compute.manager [None req-73538a34-1b2f-4a0d-b206-72fe3e124000 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61987) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 842.441028] env[61987]: DEBUG nova.objects.instance [None req-73538a34-1b2f-4a0d-b206-72fe3e124000 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lazy-loading 'flavor' on Instance uuid ad40937f-d5a9-4708-8b81-06087d38d765 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 842.471741] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61824a6d-43d5-4e4d-a185-6a88084e720f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.483029] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7dcc7a1-bc67-40ce-b6fb-87db9dffd9a1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.500980] env[61987]: DEBUG nova.compute.provider_tree [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.581560] env[61987]: DEBUG oslo_vmware.api [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061776, 'name': PowerOffVM_Task, 'duration_secs': 0.395714} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.581879] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 842.582166] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 842.582447] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-74e92f5d-0621-4742-b835-4d28a317692d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.612116] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061777, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.668382] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 842.668629] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 842.668837] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Deleting the datastore file [datastore2] 1961c2c7-4400-4559-9f7c-bec8e8e263b0 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 842.669236] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5a86fdc8-939c-49af-be64-649ccae6d415 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.671657] env[61987]: DEBUG oslo_concurrency.lockutils [req-3afec384-6dae-4d41-a08c-fca3bfa4562b req-d94d99d5-c486-4b83-915e-ad95d00efc05 service nova] Releasing lock "refresh_cache-742297eb-c45d-4f45-85d2-e9bdd3106d1e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.676666] env[61987]: DEBUG oslo_vmware.api [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 842.676666] env[61987]: value = "task-1061779" [ 842.676666] env[61987]: _type = "Task" [ 842.676666] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.685556] env[61987]: DEBUG oslo_vmware.api [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061779, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.802011] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061774, 'name': CloneVM_Task} progress is 95%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.884517] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061771, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.920504} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.884517] env[61987]: INFO nova.virt.vmwareapi.ds_util [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_1d86b010-a08b-40ca-baa2-3164788f5912/OSTACK_IMG_1d86b010-a08b-40ca-baa2-3164788f5912.vmdk to [datastore1] devstack-image-cache_base/92117f13-9c42-454c-87bb-8c7e90801e79/92117f13-9c42-454c-87bb-8c7e90801e79.vmdk. [ 842.884517] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Cleaning up location [datastore1] OSTACK_IMG_1d86b010-a08b-40ca-baa2-3164788f5912 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 842.884517] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_1d86b010-a08b-40ca-baa2-3164788f5912 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 842.884966] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-943e64ce-da02-4230-a3d9-a9571001d56d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.891078] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 842.891078] env[61987]: value = "task-1061780" [ 842.891078] env[61987]: _type = "Task" [ 842.891078] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.901693] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061780, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.910542] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5206f881-fc99-21bc-36a5-5fe15f0afaf8, 'name': SearchDatastore_Task, 'duration_secs': 0.083559} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.910889] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.911169] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 842.911424] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.911576] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.911779] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 842.912094] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a0fdebd4-79e2-4d66-a972-11f4bae60bcb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.919505] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 842.919714] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 842.920484] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2781564a-cbf1-4873-9f05-bdb9fec8780a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.925720] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 842.925720] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52148469-65a4-4979-32e7-d2baf1b968f3" [ 842.925720] env[61987]: _type = "Task" [ 842.925720] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.934256] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52148469-65a4-4979-32e7-d2baf1b968f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.977165] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-73538a34-1b2f-4a0d-b206-72fe3e124000 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 842.977494] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-649fa5de-653f-41cd-83ea-817b7f60df3c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.984578] env[61987]: DEBUG oslo_vmware.api [None req-73538a34-1b2f-4a0d-b206-72fe3e124000 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 842.984578] env[61987]: value = "task-1061781" [ 842.984578] env[61987]: _type = "Task" [ 842.984578] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.996338] env[61987]: DEBUG oslo_vmware.api [None req-73538a34-1b2f-4a0d-b206-72fe3e124000 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061781, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.005443] env[61987]: DEBUG nova.scheduler.client.report [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 843.109728] env[61987]: DEBUG oslo_vmware.api [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061777, 'name': PowerOnVM_Task, 'duration_secs': 0.665411} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.110018] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 843.114299] env[61987]: DEBUG nova.compute.manager [None req-ee5872bd-7f77-43d0-8e15-dc4278e2c4bc tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 843.115145] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ed2232-49b6-416a-a762-41da5d2a367c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.186399] env[61987]: DEBUG oslo_vmware.api [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061779, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.202725} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.186687] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 843.186879] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 843.187074] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 843.187257] env[61987]: INFO nova.compute.manager [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Took 1.15 seconds to destroy the instance on the hypervisor. [ 843.187514] env[61987]: DEBUG oslo.service.loopingcall [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 843.187708] env[61987]: DEBUG nova.compute.manager [-] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 843.187802] env[61987]: DEBUG nova.network.neutron [-] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 843.302524] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061774, 'name': CloneVM_Task, 'duration_secs': 1.619057} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.302827] env[61987]: INFO nova.virt.vmwareapi.vmops [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Created linked-clone VM from snapshot [ 843.303612] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3331e0b-b882-4709-be0c-65ede6f19ed7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.317740] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Uploading image de67bb62-3a69-4d1e-af12-db93d4352cd3 {{(pid=61987) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 843.347373] env[61987]: DEBUG oslo_vmware.rw_handles [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 843.347373] env[61987]: value = "vm-234357" [ 843.347373] env[61987]: _type = "VirtualMachine" [ 843.347373] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 843.349158] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-15a4963f-166b-4c0b-a9bb-fed20748871a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.355341] env[61987]: DEBUG oslo_vmware.rw_handles [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lease: (returnval){ [ 843.355341] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52b334b7-c9c8-4877-489e-7799ec9af1cc" [ 843.355341] env[61987]: _type = "HttpNfcLease" [ 843.355341] env[61987]: } obtained for exporting VM: (result){ [ 843.355341] env[61987]: value = "vm-234357" [ 843.355341] env[61987]: _type = "VirtualMachine" [ 843.355341] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 843.357019] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the lease: (returnval){ [ 843.357019] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52b334b7-c9c8-4877-489e-7799ec9af1cc" [ 843.357019] env[61987]: _type = "HttpNfcLease" [ 843.357019] env[61987]: } to be ready. {{(pid=61987) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 843.363797] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 843.363797] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52b334b7-c9c8-4877-489e-7799ec9af1cc" [ 843.363797] env[61987]: _type = "HttpNfcLease" [ 843.363797] env[61987]: } is initializing. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 843.401873] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061780, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.050648} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.402529] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 843.402934] env[61987]: DEBUG oslo_concurrency.lockutils [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Releasing lock "[datastore1] devstack-image-cache_base/92117f13-9c42-454c-87bb-8c7e90801e79/92117f13-9c42-454c-87bb-8c7e90801e79.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.403898] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/92117f13-9c42-454c-87bb-8c7e90801e79/92117f13-9c42-454c-87bb-8c7e90801e79.vmdk to [datastore1] 8f9aab3e-96ee-44c6-8d25-706a0ff857c4/8f9aab3e-96ee-44c6-8d25-706a0ff857c4.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 843.404874] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6ef4e62e-0c12-4ec5-a9e4-cacf534ef5a4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.413117] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 843.413117] env[61987]: value = "task-1061783" [ 843.413117] env[61987]: _type = "Task" [ 843.413117] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.423619] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061783, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.435017] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52148469-65a4-4979-32e7-d2baf1b968f3, 'name': SearchDatastore_Task, 'duration_secs': 0.012805} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.435982] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c042e558-5bc6-4ec1-8713-49cbcc42a28e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.441649] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 843.441649] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5270ef82-b38c-5a01-4c33-76bc28a62e6a" [ 843.441649] env[61987]: _type = "Task" [ 843.441649] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.445975] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 843.446129] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-11b700e9-72a9-49cb-9286-887ae13bc220 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.456540] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5270ef82-b38c-5a01-4c33-76bc28a62e6a, 'name': SearchDatastore_Task} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.458434] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.458434] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 742297eb-c45d-4f45-85d2-e9bdd3106d1e/742297eb-c45d-4f45-85d2-e9bdd3106d1e.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 843.458676] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 843.458676] env[61987]: value = "task-1061784" [ 843.458676] env[61987]: _type = "Task" [ 843.458676] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.458843] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7c2eaba7-055a-4f74-aa84-bee64aa54969 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.470103] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061784, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.471099] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 843.471099] env[61987]: value = "task-1061785" [ 843.471099] env[61987]: _type = "Task" [ 843.471099] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.481597] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061785, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.496532] env[61987]: DEBUG oslo_vmware.api [None req-73538a34-1b2f-4a0d-b206-72fe3e124000 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061781, 'name': PowerOffVM_Task, 'duration_secs': 0.388264} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.496532] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-73538a34-1b2f-4a0d-b206-72fe3e124000 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 843.496939] env[61987]: DEBUG nova.compute.manager [None req-73538a34-1b2f-4a0d-b206-72fe3e124000 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 843.497744] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e902f2f-9f9b-4ed7-bc48-b50126ac4736 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.510579] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.910s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.513101] env[61987]: DEBUG oslo_concurrency.lockutils [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.385s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.513499] env[61987]: DEBUG nova.objects.instance [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Lazy-loading 'resources' on Instance uuid 59d34772-26a5-4af2-82dc-da932e996745 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 843.556189] env[61987]: INFO nova.scheduler.client.report [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Deleted allocations for instance 4edd0159-6db1-41bd-a616-31a34e984059 [ 843.573076] env[61987]: DEBUG nova.compute.manager [req-f02ad16c-a4bd-4ad0-83f1-bfae63efcd30 req-79d2f0c2-ea21-4c48-95d8-286ac9386ea7 service nova] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Received event network-vif-deleted-6d65f8d7-1b65-4aa2-93f5-60b85af049c1 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 843.573301] env[61987]: INFO nova.compute.manager [req-f02ad16c-a4bd-4ad0-83f1-bfae63efcd30 req-79d2f0c2-ea21-4c48-95d8-286ac9386ea7 service nova] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Neutron deleted interface 6d65f8d7-1b65-4aa2-93f5-60b85af049c1; detaching it from the instance and deleting it from the info cache [ 843.573485] env[61987]: DEBUG nova.network.neutron [req-f02ad16c-a4bd-4ad0-83f1-bfae63efcd30 req-79d2f0c2-ea21-4c48-95d8-286ac9386ea7 service nova] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.867044] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 843.867044] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52b334b7-c9c8-4877-489e-7799ec9af1cc" [ 843.867044] env[61987]: _type = "HttpNfcLease" [ 843.867044] env[61987]: } is ready. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 843.867044] env[61987]: DEBUG oslo_vmware.rw_handles [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 843.867044] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52b334b7-c9c8-4877-489e-7799ec9af1cc" [ 843.867044] env[61987]: _type = "HttpNfcLease" [ 843.867044] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 843.867463] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a23b57bc-637d-471d-88ed-e0dd3bf15e25 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.875906] env[61987]: DEBUG oslo_vmware.rw_handles [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e922c6-f48a-47cc-e1ff-a5e4d6628fa4/disk-0.vmdk from lease info. {{(pid=61987) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 843.876133] env[61987]: DEBUG oslo_vmware.rw_handles [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e922c6-f48a-47cc-e1ff-a5e4d6628fa4/disk-0.vmdk for reading. {{(pid=61987) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 843.955092] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061783, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.972377] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061784, 'name': PowerOffVM_Task, 'duration_secs': 0.190071} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.975755] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 843.976624] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-474bd28a-be3d-4e90-bfc6-af7643d7f30b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.003839] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7db056cf-1858-457a-ada3-2b3b1a5f44ed {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.006925] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061785, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.007602] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f71186e9-24a3-4b88-803b-4076c76f415d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.013277] env[61987]: DEBUG oslo_concurrency.lockutils [None req-73538a34-1b2f-4a0d-b206-72fe3e124000 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ad40937f-d5a9-4708-8b81-06087d38d765" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.591s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.015201] env[61987]: DEBUG nova.network.neutron [-] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.083572] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3b4cd6e0-33a8-4ce9-a629-9d6e241a3861 tempest-FloatingIPsAssociationTestJSON-1326714254 tempest-FloatingIPsAssociationTestJSON-1326714254-project-member] Lock "4edd0159-6db1-41bd-a616-31a34e984059" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.868s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.084646] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dd106ec3-b81c-40ce-9211-4966d20a96ef {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.096315] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ed931d1-c7dc-4e5f-9f79-725412b79239 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.131488] env[61987]: DEBUG nova.compute.manager [req-f02ad16c-a4bd-4ad0-83f1-bfae63efcd30 req-79d2f0c2-ea21-4c48-95d8-286ac9386ea7 service nova] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Detach interface failed, port_id=6d65f8d7-1b65-4aa2-93f5-60b85af049c1, reason: Instance 1961c2c7-4400-4559-9f7c-bec8e8e263b0 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 844.369043] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9bae2e2-f645-48e4-8d55-9854457fa536 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.378795] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60a178ff-165b-4070-80d0-92aca5dde23f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.412015] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d844e0b0-ef1e-4e32-b74a-e29e99d5227d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.420359] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6adb204-d379-48fc-ad4d-d2d24c828490 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.435444] env[61987]: DEBUG nova.compute.provider_tree [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.446583] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061783, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.482263] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061785, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.517853] env[61987]: INFO nova.compute.manager [-] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Took 1.33 seconds to deallocate network for instance. [ 844.528024] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Creating Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 844.528024] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c9efce01-bd0c-42c7-9a2d-a2a27b7cf4ed {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.535695] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 844.535695] env[61987]: value = "task-1061786" [ 844.535695] env[61987]: _type = "Task" [ 844.535695] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.549786] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061786, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.948766] env[61987]: DEBUG nova.scheduler.client.report [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 844.966330] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061783, 'name': CopyVirtualDisk_Task} progress is 54%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.987080] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061785, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.374037} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.987503] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 742297eb-c45d-4f45-85d2-e9bdd3106d1e/742297eb-c45d-4f45-85d2-e9bdd3106d1e.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 844.987826] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 844.988232] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-644217ac-25c9-474e-b34a-115ad413911a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.996966] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 844.996966] env[61987]: value = "task-1061787" [ 844.996966] env[61987]: _type = "Task" [ 844.996966] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.012097] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061787, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.026542] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.047879] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061786, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.433044] env[61987]: DEBUG nova.objects.instance [None req-d786ef43-6941-4291-ae71-b35b0be510d1 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lazy-loading 'flavor' on Instance uuid ad40937f-d5a9-4708-8b81-06087d38d765 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 845.451025] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061783, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.459349] env[61987]: DEBUG oslo_concurrency.lockutils [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.946s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.463901] env[61987]: DEBUG oslo_concurrency.lockutils [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.928s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.464730] env[61987]: INFO nova.compute.claims [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.499279] env[61987]: INFO nova.scheduler.client.report [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Deleted allocations for instance 59d34772-26a5-4af2-82dc-da932e996745 [ 845.518535] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061787, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088199} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.518773] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 845.519675] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d72a50-3e16-4053-8dbe-57e71f0e4755 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.549812] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 742297eb-c45d-4f45-85d2-e9bdd3106d1e/742297eb-c45d-4f45-85d2-e9bdd3106d1e.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 845.549812] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f951ad50-5236-42b2-b7fb-f798c4774ad6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.577699] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 845.577699] env[61987]: value = "task-1061788" [ 845.577699] env[61987]: _type = "Task" [ 845.577699] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.582724] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061786, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.595908] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061788, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.800769] env[61987]: DEBUG nova.compute.manager [req-6a0b6c2e-67cb-4ea5-85a9-988415e41680 req-1ed908c0-7966-4abb-b3fa-6f93cf432c8e service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Received event network-changed-69ce8bb3-dc28-422e-96a8-833062bd6b77 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 845.801020] env[61987]: DEBUG nova.compute.manager [req-6a0b6c2e-67cb-4ea5-85a9-988415e41680 req-1ed908c0-7966-4abb-b3fa-6f93cf432c8e service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Refreshing instance network info cache due to event network-changed-69ce8bb3-dc28-422e-96a8-833062bd6b77. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 845.803344] env[61987]: DEBUG oslo_concurrency.lockutils [req-6a0b6c2e-67cb-4ea5-85a9-988415e41680 req-1ed908c0-7966-4abb-b3fa-6f93cf432c8e service nova] Acquiring lock "refresh_cache-98d895ee-1d1f-49cd-a055-fe6c8462a692" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.803344] env[61987]: DEBUG oslo_concurrency.lockutils [req-6a0b6c2e-67cb-4ea5-85a9-988415e41680 req-1ed908c0-7966-4abb-b3fa-6f93cf432c8e service nova] Acquired lock "refresh_cache-98d895ee-1d1f-49cd-a055-fe6c8462a692" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.803344] env[61987]: DEBUG nova.network.neutron [req-6a0b6c2e-67cb-4ea5-85a9-988415e41680 req-1ed908c0-7966-4abb-b3fa-6f93cf432c8e service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Refreshing network info cache for port 69ce8bb3-dc28-422e-96a8-833062bd6b77 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 845.938951] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d786ef43-6941-4291-ae71-b35b0be510d1 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "refresh_cache-ad40937f-d5a9-4708-8b81-06087d38d765" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.939178] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d786ef43-6941-4291-ae71-b35b0be510d1 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquired lock "refresh_cache-ad40937f-d5a9-4708-8b81-06087d38d765" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.939346] env[61987]: DEBUG nova.network.neutron [None req-d786ef43-6941-4291-ae71-b35b0be510d1 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 845.939535] env[61987]: DEBUG nova.objects.instance [None req-d786ef43-6941-4291-ae71-b35b0be510d1 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lazy-loading 'info_cache' on Instance uuid ad40937f-d5a9-4708-8b81-06087d38d765 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 845.958610] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061783, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.004427] env[61987]: DEBUG nova.compute.manager [req-a98b248b-fb96-4c33-b745-f19eb0621fc8 req-fd902d8a-c5fc-4c43-b8dc-d5c69d52a156 service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Received event network-changed-69ce8bb3-dc28-422e-96a8-833062bd6b77 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 846.004649] env[61987]: DEBUG nova.compute.manager [req-a98b248b-fb96-4c33-b745-f19eb0621fc8 req-fd902d8a-c5fc-4c43-b8dc-d5c69d52a156 service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Refreshing instance network info cache due to event network-changed-69ce8bb3-dc28-422e-96a8-833062bd6b77. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 846.004850] env[61987]: DEBUG oslo_concurrency.lockutils [req-a98b248b-fb96-4c33-b745-f19eb0621fc8 req-fd902d8a-c5fc-4c43-b8dc-d5c69d52a156 service nova] Acquiring lock "refresh_cache-98d895ee-1d1f-49cd-a055-fe6c8462a692" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.011573] env[61987]: DEBUG oslo_concurrency.lockutils [None req-578f5f15-8f69-4c6b-b687-3b606b9d93b6 tempest-ServerShowV257Test-1347971342 tempest-ServerShowV257Test-1347971342-project-member] Lock "59d34772-26a5-4af2-82dc-da932e996745" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.572s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.060417] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061786, 'name': CreateSnapshot_Task, 'duration_secs': 1.300358} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.061801] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Created Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 846.061801] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600f3e35-826d-4e34-8b8a-81b36fbb27c8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.095520] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061788, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.453874] env[61987]: DEBUG nova.objects.base [None req-d786ef43-6941-4291-ae71-b35b0be510d1 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61987) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 846.456075] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061783, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.676539} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.460198] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/92117f13-9c42-454c-87bb-8c7e90801e79/92117f13-9c42-454c-87bb-8c7e90801e79.vmdk to [datastore1] 8f9aab3e-96ee-44c6-8d25-706a0ff857c4/8f9aab3e-96ee-44c6-8d25-706a0ff857c4.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 846.461906] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ed1166-d4d6-41a4-85e0-8a50158653b6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.488548] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 8f9aab3e-96ee-44c6-8d25-706a0ff857c4/8f9aab3e-96ee-44c6-8d25-706a0ff857c4.vmdk or device None with type streamOptimized {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.495303] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8ad15846-40c7-4e92-867d-d013fff50567 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.517987] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 846.517987] env[61987]: value = "task-1061789" [ 846.517987] env[61987]: _type = "Task" [ 846.517987] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.528226] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061789, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.587411] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Creating linked-clone VM from snapshot {{(pid=61987) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 846.587411] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-be769903-5e9d-45bf-aca4-ec865fb86f1b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.604228] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061788, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.604682] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 846.604682] env[61987]: value = "task-1061790" [ 846.604682] env[61987]: _type = "Task" [ 846.604682] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.613342] env[61987]: DEBUG nova.network.neutron [req-6a0b6c2e-67cb-4ea5-85a9-988415e41680 req-1ed908c0-7966-4abb-b3fa-6f93cf432c8e service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Updated VIF entry in instance network info cache for port 69ce8bb3-dc28-422e-96a8-833062bd6b77. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 846.613835] env[61987]: DEBUG nova.network.neutron [req-6a0b6c2e-67cb-4ea5-85a9-988415e41680 req-1ed908c0-7966-4abb-b3fa-6f93cf432c8e service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Updating instance_info_cache with network_info: [{"id": "69ce8bb3-dc28-422e-96a8-833062bd6b77", "address": "fa:16:3e:ab:10:c2", "network": {"id": "35162a8d-60d3-465b-9330-64bbd235d6a2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-981766067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8dffc4d03e514fb8932b80718375084b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69ce8bb3-dc", "ovs_interfaceid": "69ce8bb3-dc28-422e-96a8-833062bd6b77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.622104] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061790, 'name': CloneVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.836029] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f8230de-e24a-44ba-af2c-32955ebbb40b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.843853] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96429ffb-b220-4756-8701-c49486dd60a8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.880361] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "interface-1f29d5c1-bd01-47dc-9dcf-4e89791120d0-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.880693] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "interface-1f29d5c1-bd01-47dc-9dcf-4e89791120d0-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.881645] env[61987]: DEBUG nova.objects.instance [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lazy-loading 'flavor' on Instance uuid 1f29d5c1-bd01-47dc-9dcf-4e89791120d0 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 846.883802] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88cf6e0b-91e8-4d4e-a196-39d3d49ff154 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.893996] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4b7e4c5-6b2a-4488-a3b4-ee6516f300ca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.910804] env[61987]: DEBUG nova.compute.provider_tree [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.031030] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061789, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.096668] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061788, 'name': ReconfigVM_Task, 'duration_secs': 1.136279} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.099341] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 742297eb-c45d-4f45-85d2-e9bdd3106d1e/742297eb-c45d-4f45-85d2-e9bdd3106d1e.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.100040] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-032dcd1c-da12-4f74-ab76-473efdabc6b9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.110784] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 847.110784] env[61987]: value = "task-1061791" [ 847.110784] env[61987]: _type = "Task" [ 847.110784] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.117685] env[61987]: DEBUG oslo_concurrency.lockutils [req-6a0b6c2e-67cb-4ea5-85a9-988415e41680 req-1ed908c0-7966-4abb-b3fa-6f93cf432c8e service nova] Releasing lock "refresh_cache-98d895ee-1d1f-49cd-a055-fe6c8462a692" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.118639] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061790, 'name': CloneVM_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.119756] env[61987]: DEBUG oslo_concurrency.lockutils [req-a98b248b-fb96-4c33-b745-f19eb0621fc8 req-fd902d8a-c5fc-4c43-b8dc-d5c69d52a156 service nova] Acquired lock "refresh_cache-98d895ee-1d1f-49cd-a055-fe6c8462a692" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.119756] env[61987]: DEBUG nova.network.neutron [req-a98b248b-fb96-4c33-b745-f19eb0621fc8 req-fd902d8a-c5fc-4c43-b8dc-d5c69d52a156 service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Refreshing network info cache for port 69ce8bb3-dc28-422e-96a8-833062bd6b77 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 847.124222] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061791, 'name': Rename_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.250491] env[61987]: DEBUG nova.network.neutron [None req-d786ef43-6941-4291-ae71-b35b0be510d1 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Updating instance_info_cache with network_info: [{"id": "6aa03b99-d58f-461e-b2a8-c4552c1d418a", "address": "fa:16:3e:5e:3e:15", "network": {"id": "a9093c43-6bc4-45b1-bdcf-a434759300bd", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-355385599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.128", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7efd6c745f547d3803d5a6184547bda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac563aa7-6d7c-4bd5-9241-7b3e11b8f22d", "external-id": "nsx-vlan-transportzone-730", "segmentation_id": 730, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6aa03b99-d5", "ovs_interfaceid": "6aa03b99-d58f-461e-b2a8-c4552c1d418a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.416349] env[61987]: DEBUG nova.scheduler.client.report [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 847.478086] env[61987]: DEBUG nova.objects.instance [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lazy-loading 'pci_requests' on Instance uuid 1f29d5c1-bd01-47dc-9dcf-4e89791120d0 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 847.530287] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061789, 'name': ReconfigVM_Task, 'duration_secs': 0.832334} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.530715] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 8f9aab3e-96ee-44c6-8d25-706a0ff857c4/8f9aab3e-96ee-44c6-8d25-706a0ff857c4.vmdk or device None with type streamOptimized {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.531332] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1c48e6cf-4724-4be1-b19e-cfda1d97b062 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.537672] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 847.537672] env[61987]: value = "task-1061792" [ 847.537672] env[61987]: _type = "Task" [ 847.537672] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.549159] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061792, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.620162] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061790, 'name': CloneVM_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.625850] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061791, 'name': Rename_Task, 'duration_secs': 0.267611} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.626164] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 847.626489] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-da261b4d-cbd4-443b-8cd1-5ca9eaaed380 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.635851] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 847.635851] env[61987]: value = "task-1061793" [ 847.635851] env[61987]: _type = "Task" [ 847.635851] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.647285] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061793, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.753520] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d786ef43-6941-4291-ae71-b35b0be510d1 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Releasing lock "refresh_cache-ad40937f-d5a9-4708-8b81-06087d38d765" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.883783] env[61987]: DEBUG nova.network.neutron [req-a98b248b-fb96-4c33-b745-f19eb0621fc8 req-fd902d8a-c5fc-4c43-b8dc-d5c69d52a156 service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Updated VIF entry in instance network info cache for port 69ce8bb3-dc28-422e-96a8-833062bd6b77. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 847.884437] env[61987]: DEBUG nova.network.neutron [req-a98b248b-fb96-4c33-b745-f19eb0621fc8 req-fd902d8a-c5fc-4c43-b8dc-d5c69d52a156 service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Updating instance_info_cache with network_info: [{"id": "69ce8bb3-dc28-422e-96a8-833062bd6b77", "address": "fa:16:3e:ab:10:c2", "network": {"id": "35162a8d-60d3-465b-9330-64bbd235d6a2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-981766067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8dffc4d03e514fb8932b80718375084b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69ce8bb3-dc", "ovs_interfaceid": "69ce8bb3-dc28-422e-96a8-833062bd6b77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.924548] env[61987]: DEBUG oslo_concurrency.lockutils [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.461s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.924548] env[61987]: DEBUG nova.compute.manager [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 847.927584] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.807s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.927584] env[61987]: DEBUG nova.objects.instance [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Lazy-loading 'resources' on Instance uuid 6ed3f0e9-3c39-4406-9863-931a117961a8 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 847.980913] env[61987]: DEBUG nova.objects.base [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Object Instance<1f29d5c1-bd01-47dc-9dcf-4e89791120d0> lazy-loaded attributes: flavor,pci_requests {{(pid=61987) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 847.981390] env[61987]: DEBUG nova.network.neutron [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 848.026325] env[61987]: DEBUG nova.policy [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '697f9b1e3b3840369549a4f1ae3173ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd2afdc9bddba4ccab2b8fdb1f31da2b6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 848.051481] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061792, 'name': Rename_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.123714] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061790, 'name': CloneVM_Task} progress is 95%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.148061] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061793, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.313083] env[61987]: DEBUG nova.network.neutron [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Successfully created port: ebd9819d-a92d-4034-a00b-af37dbb338ab {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 848.387308] env[61987]: DEBUG oslo_concurrency.lockutils [req-a98b248b-fb96-4c33-b745-f19eb0621fc8 req-fd902d8a-c5fc-4c43-b8dc-d5c69d52a156 service nova] Releasing lock "refresh_cache-98d895ee-1d1f-49cd-a055-fe6c8462a692" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.430108] env[61987]: DEBUG nova.compute.utils [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 848.439263] env[61987]: DEBUG nova.compute.manager [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 848.439461] env[61987]: DEBUG nova.network.neutron [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 848.510064] env[61987]: DEBUG nova.policy [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4ead7955f1d475590523ac98c579956', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '315a2a3450024ed38e27cd7758f392f9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 848.557789] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061792, 'name': Rename_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.620835] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061790, 'name': CloneVM_Task, 'duration_secs': 1.630222} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.621582] env[61987]: INFO nova.virt.vmwareapi.vmops [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Created linked-clone VM from snapshot [ 848.621928] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eeea153-158c-417e-a923-1774133010b4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.633929] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Uploading image 87a46110-ae87-47e9-975e-48f434a17e9a {{(pid=61987) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 848.653291] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061793, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.669178] env[61987]: DEBUG oslo_vmware.rw_handles [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 848.669178] env[61987]: value = "vm-234360" [ 848.669178] env[61987]: _type = "VirtualMachine" [ 848.669178] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 848.669528] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-15033d87-6714-405d-b1b8-87ca0cbe8f29 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.677670] env[61987]: DEBUG oslo_vmware.rw_handles [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lease: (returnval){ [ 848.677670] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]524edabc-03bd-e58e-2292-d2a308cf4486" [ 848.677670] env[61987]: _type = "HttpNfcLease" [ 848.677670] env[61987]: } obtained for exporting VM: (result){ [ 848.677670] env[61987]: value = "vm-234360" [ 848.677670] env[61987]: _type = "VirtualMachine" [ 848.677670] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 848.677939] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the lease: (returnval){ [ 848.677939] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]524edabc-03bd-e58e-2292-d2a308cf4486" [ 848.677939] env[61987]: _type = "HttpNfcLease" [ 848.677939] env[61987]: } to be ready. {{(pid=61987) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 848.685378] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 848.685378] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]524edabc-03bd-e58e-2292-d2a308cf4486" [ 848.685378] env[61987]: _type = "HttpNfcLease" [ 848.685378] env[61987]: } is initializing. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 848.760328] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d786ef43-6941-4291-ae71-b35b0be510d1 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 848.761460] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-000c3da7-4ef6-4c3d-bef8-8d6a0becef79 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.765743] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-278a8830-9a1a-442e-957d-da0924560460 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.774250] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4440fc8c-4412-4f9d-8abd-3d0c1770a553 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.778602] env[61987]: DEBUG oslo_vmware.api [None req-d786ef43-6941-4291-ae71-b35b0be510d1 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 848.778602] env[61987]: value = "task-1061795" [ 848.778602] env[61987]: _type = "Task" [ 848.778602] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.816732] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0658f54b-f32f-48c9-99c4-f462bdf9578d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.824337] env[61987]: DEBUG oslo_vmware.api [None req-d786ef43-6941-4291-ae71-b35b0be510d1 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061795, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.829995] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3db528f-67e9-4271-a580-f57dcb66a1d0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.845028] env[61987]: DEBUG nova.compute.provider_tree [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.939940] env[61987]: DEBUG nova.compute.manager [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 848.962096] env[61987]: DEBUG nova.network.neutron [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Successfully created port: 507c557c-1a2c-4bcb-a8da-93fc1fa91bd9 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 849.052186] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061792, 'name': Rename_Task, 'duration_secs': 1.182286} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.052738] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 849.053126] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9ac8279a-0b4f-48a6-9648-df733ae40c12 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.065037] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 849.065037] env[61987]: value = "task-1061796" [ 849.065037] env[61987]: _type = "Task" [ 849.065037] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.073576] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061796, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.148948] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061793, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.188943] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 849.188943] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]524edabc-03bd-e58e-2292-d2a308cf4486" [ 849.188943] env[61987]: _type = "HttpNfcLease" [ 849.188943] env[61987]: } is ready. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 849.188943] env[61987]: DEBUG oslo_vmware.rw_handles [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 849.188943] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]524edabc-03bd-e58e-2292-d2a308cf4486" [ 849.188943] env[61987]: _type = "HttpNfcLease" [ 849.188943] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 849.193019] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e10ea0a-3a77-4a4a-a1e8-215e23170816 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.197764] env[61987]: DEBUG oslo_vmware.rw_handles [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52485f60-c3dd-e925-b0d2-ce13f8afb9c7/disk-0.vmdk from lease info. {{(pid=61987) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 849.197966] env[61987]: DEBUG oslo_vmware.rw_handles [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52485f60-c3dd-e925-b0d2-ce13f8afb9c7/disk-0.vmdk for reading. {{(pid=61987) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 849.291774] env[61987]: DEBUG oslo_vmware.api [None req-d786ef43-6941-4291-ae71-b35b0be510d1 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061795, 'name': PowerOnVM_Task, 'duration_secs': 0.405098} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.292283] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d786ef43-6941-4291-ae71-b35b0be510d1 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 849.292627] env[61987]: DEBUG nova.compute.manager [None req-d786ef43-6941-4291-ae71-b35b0be510d1 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 849.293889] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8318a89-3e96-4b4f-84ea-3523d3730333 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.345514] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ac6cdc9c-6033-4a08-8478-31c1aca29be3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.348955] env[61987]: DEBUG nova.scheduler.client.report [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 849.576771] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061796, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.650989] env[61987]: DEBUG oslo_vmware.api [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061793, 'name': PowerOnVM_Task, 'duration_secs': 1.759429} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.651322] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 849.651560] env[61987]: INFO nova.compute.manager [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Took 11.00 seconds to spawn the instance on the hypervisor. [ 849.651753] env[61987]: DEBUG nova.compute.manager [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 849.652588] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fd70bae-82b4-4d05-b6fa-f4e97bb21157 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.856234] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.927s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.857801] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.611s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.858268] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.860911] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.277s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.862967] env[61987]: INFO nova.compute.claims [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 849.889125] env[61987]: DEBUG nova.network.neutron [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Successfully updated port: ebd9819d-a92d-4034-a00b-af37dbb338ab {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 849.892079] env[61987]: INFO nova.scheduler.client.report [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Deleted allocations for instance 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2 [ 849.896869] env[61987]: INFO nova.scheduler.client.report [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Deleted allocations for instance 6ed3f0e9-3c39-4406-9863-931a117961a8 [ 849.951490] env[61987]: DEBUG nova.compute.manager [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 850.077566] env[61987]: DEBUG oslo_vmware.api [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061796, 'name': PowerOnVM_Task, 'duration_secs': 0.658266} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.077861] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 850.078201] env[61987]: INFO nova.compute.manager [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Took 16.94 seconds to spawn the instance on the hypervisor. [ 850.078376] env[61987]: DEBUG nova.compute.manager [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 850.080864] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7ea4f58-d248-4dca-a352-bb9c356b5f2d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.176034] env[61987]: INFO nova.compute.manager [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Took 35.25 seconds to build instance. [ 850.249715] env[61987]: DEBUG nova.compute.manager [req-2e601eaa-369d-4f28-84ef-be49486358fd req-2037e2ec-a29e-42ed-8133-e42e2592db8e service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Received event network-changed-69ce8bb3-dc28-422e-96a8-833062bd6b77 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 850.250075] env[61987]: DEBUG nova.compute.manager [req-2e601eaa-369d-4f28-84ef-be49486358fd req-2037e2ec-a29e-42ed-8133-e42e2592db8e service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Refreshing instance network info cache due to event network-changed-69ce8bb3-dc28-422e-96a8-833062bd6b77. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 850.250456] env[61987]: DEBUG oslo_concurrency.lockutils [req-2e601eaa-369d-4f28-84ef-be49486358fd req-2037e2ec-a29e-42ed-8133-e42e2592db8e service nova] Acquiring lock "refresh_cache-98d895ee-1d1f-49cd-a055-fe6c8462a692" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.250691] env[61987]: DEBUG oslo_concurrency.lockutils [req-2e601eaa-369d-4f28-84ef-be49486358fd req-2037e2ec-a29e-42ed-8133-e42e2592db8e service nova] Acquired lock "refresh_cache-98d895ee-1d1f-49cd-a055-fe6c8462a692" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.250908] env[61987]: DEBUG nova.network.neutron [req-2e601eaa-369d-4f28-84ef-be49486358fd req-2037e2ec-a29e-42ed-8133-e42e2592db8e service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Refreshing network info cache for port 69ce8bb3-dc28-422e-96a8-833062bd6b77 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 850.395725] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "refresh_cache-1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.395924] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "refresh_cache-1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.396091] env[61987]: DEBUG nova.network.neutron [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 850.409231] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5fc4a2e0-433d-4e09-9884-d32cf8858a42 tempest-InstanceActionsV221TestJSON-1146305562 tempest-InstanceActionsV221TestJSON-1146305562-project-member] Lock "6ed3f0e9-3c39-4406-9863-931a117961a8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.818s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.410632] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1f067276-395f-491e-bc8f-7e24c1e39b0a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.281s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.602496] env[61987]: INFO nova.compute.manager [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Took 44.10 seconds to build instance. [ 850.676899] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6545aa20-009f-4713-81ab-7e79412cc6f5 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "742297eb-c45d-4f45-85d2-e9bdd3106d1e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.768s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.929170] env[61987]: DEBUG nova.network.neutron [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Successfully updated port: 507c557c-1a2c-4bcb-a8da-93fc1fa91bd9 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 851.088446] env[61987]: WARNING nova.network.neutron [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] 257456d2-c9db-4303-bdfd-035bbb123901 already exists in list: networks containing: ['257456d2-c9db-4303-bdfd-035bbb123901']. ignoring it [ 851.104473] env[61987]: DEBUG oslo_concurrency.lockutils [None req-687b4b27-fa70-46a2-af4e-aaffe058fd59 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "8f9aab3e-96ee-44c6-8d25-706a0ff857c4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.625s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.129119] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17658d47-a6c3-4bc6-af39-c749836be8f6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.137691] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8232bc-5417-4468-8579-95a1ac0327c0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.177135] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd18098-6bb9-4e8e-85bf-2ab65ff41b37 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.185681] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e0be97-5299-41d4-9763-cc274aad02fc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.201401] env[61987]: DEBUG nova.compute.provider_tree [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.342281] env[61987]: DEBUG nova.network.neutron [req-2e601eaa-369d-4f28-84ef-be49486358fd req-2037e2ec-a29e-42ed-8133-e42e2592db8e service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Updated VIF entry in instance network info cache for port 69ce8bb3-dc28-422e-96a8-833062bd6b77. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 851.342771] env[61987]: DEBUG nova.network.neutron [req-2e601eaa-369d-4f28-84ef-be49486358fd req-2037e2ec-a29e-42ed-8133-e42e2592db8e service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Updating instance_info_cache with network_info: [{"id": "69ce8bb3-dc28-422e-96a8-833062bd6b77", "address": "fa:16:3e:ab:10:c2", "network": {"id": "35162a8d-60d3-465b-9330-64bbd235d6a2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-981766067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8dffc4d03e514fb8932b80718375084b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69ce8bb3-dc", "ovs_interfaceid": "69ce8bb3-dc28-422e-96a8-833062bd6b77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.432704] env[61987]: DEBUG oslo_concurrency.lockutils [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "refresh_cache-ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.432976] env[61987]: DEBUG oslo_concurrency.lockutils [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired lock "refresh_cache-ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.433068] env[61987]: DEBUG nova.network.neutron [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 851.525944] env[61987]: DEBUG nova.network.neutron [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Updating instance_info_cache with network_info: [{"id": "e7d811c0-d77e-448d-86fe-acdb61c04284", "address": "fa:16:3e:b0:6d:8c", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7d811c0-d7", "ovs_interfaceid": "e7d811c0-d77e-448d-86fe-acdb61c04284", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "ebd9819d-a92d-4034-a00b-af37dbb338ab", "address": "fa:16:3e:36:15:13", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebd9819d-a9", "ovs_interfaceid": "ebd9819d-a92d-4034-a00b-af37dbb338ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.705260] env[61987]: DEBUG nova.scheduler.client.report [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 851.846756] env[61987]: DEBUG oslo_concurrency.lockutils [req-2e601eaa-369d-4f28-84ef-be49486358fd req-2037e2ec-a29e-42ed-8133-e42e2592db8e service nova] Releasing lock "refresh_cache-98d895ee-1d1f-49cd-a055-fe6c8462a692" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.971237] env[61987]: DEBUG nova.network.neutron [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 852.029962] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "refresh_cache-1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.030644] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.030814] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.031700] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb96b651-956b-4e5a-95c1-74bfb04ff47c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.050588] env[61987]: DEBUG nova.virt.hardware [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 852.050854] env[61987]: DEBUG nova.virt.hardware [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 852.051037] env[61987]: DEBUG nova.virt.hardware [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 852.051240] env[61987]: DEBUG nova.virt.hardware [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 852.051395] env[61987]: DEBUG nova.virt.hardware [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 852.051552] env[61987]: DEBUG nova.virt.hardware [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 852.051980] env[61987]: DEBUG nova.virt.hardware [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 852.051980] env[61987]: DEBUG nova.virt.hardware [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 852.052140] env[61987]: DEBUG nova.virt.hardware [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 852.052345] env[61987]: DEBUG nova.virt.hardware [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 852.052591] env[61987]: DEBUG nova.virt.hardware [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 852.059017] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Reconfiguring VM to attach interface {{(pid=61987) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 852.061915] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-334c1594-034a-4875-8417-60dc2a88acc9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.079984] env[61987]: DEBUG oslo_vmware.api [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 852.079984] env[61987]: value = "task-1061797" [ 852.079984] env[61987]: _type = "Task" [ 852.079984] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.088535] env[61987]: DEBUG oslo_vmware.api [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061797, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.149597] env[61987]: DEBUG nova.network.neutron [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Updating instance_info_cache with network_info: [{"id": "507c557c-1a2c-4bcb-a8da-93fc1fa91bd9", "address": "fa:16:3e:8c:d5:4a", "network": {"id": "59d8d7d6-fd71-4dec-802e-e0f9e670ff70", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1845227031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "315a2a3450024ed38e27cd7758f392f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap507c557c-1a", "ovs_interfaceid": "507c557c-1a2c-4bcb-a8da-93fc1fa91bd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.210857] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.350s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.211458] env[61987]: DEBUG nova.compute.manager [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 852.214579] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.188s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.214743] env[61987]: DEBUG nova.objects.instance [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lazy-loading 'resources' on Instance uuid 1961c2c7-4400-4559-9f7c-bec8e8e263b0 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 852.591147] env[61987]: DEBUG oslo_vmware.api [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061797, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.653548] env[61987]: DEBUG oslo_concurrency.lockutils [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Releasing lock "refresh_cache-ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.653548] env[61987]: DEBUG nova.compute.manager [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Instance network_info: |[{"id": "507c557c-1a2c-4bcb-a8da-93fc1fa91bd9", "address": "fa:16:3e:8c:d5:4a", "network": {"id": "59d8d7d6-fd71-4dec-802e-e0f9e670ff70", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1845227031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "315a2a3450024ed38e27cd7758f392f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap507c557c-1a", "ovs_interfaceid": "507c557c-1a2c-4bcb-a8da-93fc1fa91bd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 852.719069] env[61987]: DEBUG nova.compute.utils [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 852.722738] env[61987]: DEBUG nova.compute.manager [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 852.722905] env[61987]: DEBUG nova.network.neutron [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 852.770711] env[61987]: DEBUG nova.policy [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a3425a6d5e36481bb702ddf789ea9eed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0c3d31c138df453486a2b1a6b70a84d4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 852.932308] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10c6a300-17d2-4e81-a6c6-422a6f4c6fc3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.941644] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf60302-12c0-41b9-86f6-9d7d2554341a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.972956] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2eeb08c-cc0d-4114-a40c-92092aa0d6c9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.981780] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf81823-40a1-4e00-8b3c-25f7d426e850 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.996476] env[61987]: DEBUG nova.compute.provider_tree [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.050995] env[61987]: DEBUG nova.network.neutron [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Successfully created port: 1460c9a3-914a-4c65-a2c3-601f1e07788e {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 853.090207] env[61987]: DEBUG oslo_vmware.api [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061797, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.223843] env[61987]: DEBUG nova.compute.manager [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 853.500162] env[61987]: DEBUG nova.scheduler.client.report [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 853.591706] env[61987]: DEBUG oslo_vmware.api [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061797, 'name': ReconfigVM_Task, 'duration_secs': 1.123507} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.592364] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.592598] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Reconfigured VM to attach interface {{(pid=61987) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 853.696887] env[61987]: DEBUG nova.virt.hardware [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 853.697133] env[61987]: DEBUG nova.virt.hardware [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 853.697412] env[61987]: DEBUG nova.virt.hardware [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 853.697801] env[61987]: DEBUG nova.virt.hardware [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 853.698079] env[61987]: DEBUG nova.virt.hardware [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 853.698346] env[61987]: DEBUG nova.virt.hardware [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 853.698684] env[61987]: DEBUG nova.virt.hardware [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 853.698945] env[61987]: DEBUG nova.virt.hardware [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 853.699253] env[61987]: DEBUG nova.virt.hardware [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 853.699554] env[61987]: DEBUG nova.virt.hardware [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 853.699869] env[61987]: DEBUG nova.virt.hardware [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 853.702982] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11c2a143-e12f-4cee-8bf9-902636287b64 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.715217] env[61987]: DEBUG oslo_vmware.rw_handles [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e922c6-f48a-47cc-e1ff-a5e4d6628fa4/disk-0.vmdk. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 853.717058] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d1e1920-4e43-4c40-97ee-a21865a4812b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.721089] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d7ac5e-cc56-4b7d-a8cd-a5dc4f73f0f4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.738585] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8c:d5:4a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '028bae2d-fe6c-4207-b4a3-3fab45fbf1d6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '507c557c-1a2c-4bcb-a8da-93fc1fa91bd9', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 853.746011] env[61987]: DEBUG oslo.service.loopingcall [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.748096] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 853.748384] env[61987]: DEBUG oslo_vmware.rw_handles [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e922c6-f48a-47cc-e1ff-a5e4d6628fa4/disk-0.vmdk is in state: ready. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 853.748546] env[61987]: ERROR oslo_vmware.rw_handles [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e922c6-f48a-47cc-e1ff-a5e4d6628fa4/disk-0.vmdk due to incomplete transfer. [ 853.748773] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a90626cb-11d2-4fe8-a9b7-dd351740cc2c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.763829] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-a5a8aea4-265c-4364-ad0f-67d6240789e8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.772013] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 853.772013] env[61987]: value = "task-1061798" [ 853.772013] env[61987]: _type = "Task" [ 853.772013] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.773284] env[61987]: DEBUG oslo_vmware.rw_handles [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e922c6-f48a-47cc-e1ff-a5e4d6628fa4/disk-0.vmdk. {{(pid=61987) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 853.773930] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Uploaded image de67bb62-3a69-4d1e-af12-db93d4352cd3 to the Glance image server {{(pid=61987) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 853.775833] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Destroying the VM {{(pid=61987) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 853.779541] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e1bcb588-0cda-4d23-95f6-ec855ab059b2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.787704] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061798, 'name': CreateVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.789162] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 853.789162] env[61987]: value = "task-1061799" [ 853.789162] env[61987]: _type = "Task" [ 853.789162] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.797310] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061799, 'name': Destroy_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.997839] env[61987]: DEBUG nova.compute.manager [req-a6fef8cb-a80c-426b-865b-e0cdd1fdbcb4 req-242ebd72-7400-4fc9-accc-e6f5c9cceb81 service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Received event network-vif-plugged-ebd9819d-a92d-4034-a00b-af37dbb338ab {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 853.997839] env[61987]: DEBUG oslo_concurrency.lockutils [req-a6fef8cb-a80c-426b-865b-e0cdd1fdbcb4 req-242ebd72-7400-4fc9-accc-e6f5c9cceb81 service nova] Acquiring lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.997976] env[61987]: DEBUG oslo_concurrency.lockutils [req-a6fef8cb-a80c-426b-865b-e0cdd1fdbcb4 req-242ebd72-7400-4fc9-accc-e6f5c9cceb81 service nova] Lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.998173] env[61987]: DEBUG oslo_concurrency.lockutils [req-a6fef8cb-a80c-426b-865b-e0cdd1fdbcb4 req-242ebd72-7400-4fc9-accc-e6f5c9cceb81 service nova] Lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.998359] env[61987]: DEBUG nova.compute.manager [req-a6fef8cb-a80c-426b-865b-e0cdd1fdbcb4 req-242ebd72-7400-4fc9-accc-e6f5c9cceb81 service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] No waiting events found dispatching network-vif-plugged-ebd9819d-a92d-4034-a00b-af37dbb338ab {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 853.998534] env[61987]: WARNING nova.compute.manager [req-a6fef8cb-a80c-426b-865b-e0cdd1fdbcb4 req-242ebd72-7400-4fc9-accc-e6f5c9cceb81 service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Received unexpected event network-vif-plugged-ebd9819d-a92d-4034-a00b-af37dbb338ab for instance with vm_state active and task_state None. [ 854.005343] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.791s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.027190] env[61987]: INFO nova.scheduler.client.report [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Deleted allocations for instance 1961c2c7-4400-4559-9f7c-bec8e8e263b0 [ 854.098728] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fa1ad601-060f-497f-b81b-4715971ea942 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "interface-1f29d5c1-bd01-47dc-9dcf-4e89791120d0-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.217s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.240095] env[61987]: DEBUG nova.compute.manager [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 854.268533] env[61987]: DEBUG nova.virt.hardware [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 854.268860] env[61987]: DEBUG nova.virt.hardware [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 854.268988] env[61987]: DEBUG nova.virt.hardware [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 854.269198] env[61987]: DEBUG nova.virt.hardware [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 854.269354] env[61987]: DEBUG nova.virt.hardware [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 854.269538] env[61987]: DEBUG nova.virt.hardware [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 854.269758] env[61987]: DEBUG nova.virt.hardware [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 854.269925] env[61987]: DEBUG nova.virt.hardware [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 854.270128] env[61987]: DEBUG nova.virt.hardware [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 854.270309] env[61987]: DEBUG nova.virt.hardware [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 854.270496] env[61987]: DEBUG nova.virt.hardware [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 854.271418] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e20941d-0660-4c34-8ec5-755f3f1fbb22 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.287223] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e5d8965-9545-4334-92fd-7419a2e509a2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.291155] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061798, 'name': CreateVM_Task, 'duration_secs': 0.502999} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.291394] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 854.295243] env[61987]: DEBUG oslo_concurrency.lockutils [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.295531] env[61987]: DEBUG oslo_concurrency.lockutils [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.295877] env[61987]: DEBUG oslo_concurrency.lockutils [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 854.296213] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f32ed986-0e6f-49e1-a52c-a79b651dc5b0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.310938] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061799, 'name': Destroy_Task, 'duration_secs': 0.500485} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.312281] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Destroyed the VM [ 854.312556] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Deleting Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 854.312909] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 854.312909] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]524f4f55-1500-ce63-0399-57775c480543" [ 854.312909] env[61987]: _type = "Task" [ 854.312909] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.313120] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a5b30105-427b-4515-8ea2-f7e41345e7d3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.324758] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524f4f55-1500-ce63-0399-57775c480543, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.326300] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 854.326300] env[61987]: value = "task-1061800" [ 854.326300] env[61987]: _type = "Task" [ 854.326300] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.334043] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061800, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.536247] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5fa9ce2a-f349-4656-b8df-d0284fc22fd8 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "1961c2c7-4400-4559-9f7c-bec8e8e263b0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.010s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.554763] env[61987]: DEBUG nova.network.neutron [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Successfully updated port: 1460c9a3-914a-4c65-a2c3-601f1e07788e {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 854.603179] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquiring lock "455562c7-9fc1-4625-95b0-5eb8c9c8ccaa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.603333] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "455562c7-9fc1-4625-95b0-5eb8c9c8ccaa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.827149] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524f4f55-1500-ce63-0399-57775c480543, 'name': SearchDatastore_Task, 'duration_secs': 0.016132} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.830702] env[61987]: DEBUG oslo_concurrency.lockutils [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.832033] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 854.832033] env[61987]: DEBUG oslo_concurrency.lockutils [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.832033] env[61987]: DEBUG oslo_concurrency.lockutils [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.832033] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 854.832033] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-14537e5e-17d8-4924-99b7-3d3d0ff7e6f5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.839240] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061800, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.848935] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 854.848935] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 854.848935] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-244928ff-79a1-4a46-ab3d-53a8d803058c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.854848] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 854.854848] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52496385-bcf5-5800-821d-5928ab25509c" [ 854.854848] env[61987]: _type = "Task" [ 854.854848] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.864472] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52496385-bcf5-5800-821d-5928ab25509c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.973369] env[61987]: DEBUG nova.compute.manager [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Received event network-vif-plugged-507c557c-1a2c-4bcb-a8da-93fc1fa91bd9 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 854.973369] env[61987]: DEBUG oslo_concurrency.lockutils [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] Acquiring lock "ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.973369] env[61987]: DEBUG oslo_concurrency.lockutils [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] Lock "ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.973369] env[61987]: DEBUG oslo_concurrency.lockutils [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] Lock "ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.973369] env[61987]: DEBUG nova.compute.manager [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] No waiting events found dispatching network-vif-plugged-507c557c-1a2c-4bcb-a8da-93fc1fa91bd9 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 854.973369] env[61987]: WARNING nova.compute.manager [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Received unexpected event network-vif-plugged-507c557c-1a2c-4bcb-a8da-93fc1fa91bd9 for instance with vm_state building and task_state spawning. [ 854.973369] env[61987]: DEBUG nova.compute.manager [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Received event network-changed-507c557c-1a2c-4bcb-a8da-93fc1fa91bd9 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 854.973609] env[61987]: DEBUG nova.compute.manager [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Refreshing instance network info cache due to event network-changed-507c557c-1a2c-4bcb-a8da-93fc1fa91bd9. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 854.973609] env[61987]: DEBUG oslo_concurrency.lockutils [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] Acquiring lock "refresh_cache-ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.973873] env[61987]: DEBUG oslo_concurrency.lockutils [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] Acquired lock "refresh_cache-ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.973873] env[61987]: DEBUG nova.network.neutron [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Refreshing network info cache for port 507c557c-1a2c-4bcb-a8da-93fc1fa91bd9 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 855.058037] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "refresh_cache-1bba4036-5211-4a43-b2e6-5276851335e2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.058037] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquired lock "refresh_cache-1bba4036-5211-4a43-b2e6-5276851335e2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.058037] env[61987]: DEBUG nova.network.neutron [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 855.106613] env[61987]: DEBUG nova.compute.manager [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 855.337263] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061800, 'name': RemoveSnapshot_Task, 'duration_secs': 0.532085} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.337613] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Deleted Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 855.337932] env[61987]: DEBUG nova.compute.manager [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 855.338729] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f1aeeb0-7db0-4181-8ef5-62c0ca6285a6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.374022] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52496385-bcf5-5800-821d-5928ab25509c, 'name': SearchDatastore_Task, 'duration_secs': 0.013773} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.374022] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-451ff8cc-1a38-474c-97a6-f30ea1480158 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.379893] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 855.379893] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]524a9f91-be39-497b-b333-117ec79088da" [ 855.379893] env[61987]: _type = "Task" [ 855.379893] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.389152] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524a9f91-be39-497b-b333-117ec79088da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.635820] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.636125] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.637690] env[61987]: INFO nova.compute.claims [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 855.649235] env[61987]: DEBUG nova.network.neutron [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 855.715280] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "8f9aab3e-96ee-44c6-8d25-706a0ff857c4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.715604] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "8f9aab3e-96ee-44c6-8d25-706a0ff857c4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.715827] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "8f9aab3e-96ee-44c6-8d25-706a0ff857c4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.716035] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "8f9aab3e-96ee-44c6-8d25-706a0ff857c4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.716223] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "8f9aab3e-96ee-44c6-8d25-706a0ff857c4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.719087] env[61987]: INFO nova.compute.manager [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Terminating instance [ 855.860391] env[61987]: INFO nova.compute.manager [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Shelve offloading [ 855.896728] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524a9f91-be39-497b-b333-117ec79088da, 'name': SearchDatastore_Task, 'duration_secs': 0.013132} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.897061] env[61987]: DEBUG oslo_concurrency.lockutils [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.897499] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05/ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 855.897689] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-75b97911-8113-4fac-99d7-d7825f51a9b1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.906596] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 855.906596] env[61987]: value = "task-1061801" [ 855.906596] env[61987]: _type = "Task" [ 855.906596] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.915575] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061801, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.093423] env[61987]: DEBUG nova.network.neutron [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Updating instance_info_cache with network_info: [{"id": "1460c9a3-914a-4c65-a2c3-601f1e07788e", "address": "fa:16:3e:c2:75:a9", "network": {"id": "8dfc473a-6d57-4169-9338-49a9565bb93d", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-422412073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c3d31c138df453486a2b1a6b70a84d4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1460c9a3-91", "ovs_interfaceid": "1460c9a3-914a-4c65-a2c3-601f1e07788e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.106079] env[61987]: DEBUG nova.network.neutron [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Updated VIF entry in instance network info cache for port 507c557c-1a2c-4bcb-a8da-93fc1fa91bd9. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 856.106697] env[61987]: DEBUG nova.network.neutron [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Updating instance_info_cache with network_info: [{"id": "507c557c-1a2c-4bcb-a8da-93fc1fa91bd9", "address": "fa:16:3e:8c:d5:4a", "network": {"id": "59d8d7d6-fd71-4dec-802e-e0f9e670ff70", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1845227031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "315a2a3450024ed38e27cd7758f392f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap507c557c-1a", "ovs_interfaceid": "507c557c-1a2c-4bcb-a8da-93fc1fa91bd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.225028] env[61987]: DEBUG nova.compute.manager [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 856.225028] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 856.225028] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b45883d5-c940-4fd0-94d5-829410f99cb0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.235550] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 856.236019] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-353e0194-2948-4170-a421-20821947dd9b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.245974] env[61987]: DEBUG oslo_vmware.api [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 856.245974] env[61987]: value = "task-1061802" [ 856.245974] env[61987]: _type = "Task" [ 856.245974] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.258230] env[61987]: DEBUG oslo_vmware.api [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061802, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.377030] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 856.377030] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8d71026-9613-4e8b-8700-e2faf4f3f78c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.385401] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 856.385401] env[61987]: value = "task-1061803" [ 856.385401] env[61987]: _type = "Task" [ 856.385401] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.414076] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] VM already powered off {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 856.414804] env[61987]: DEBUG nova.compute.manager [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 856.417640] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff3b21e-1a2a-4764-b8ac-27659dda9e05 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.430188] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061801, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.433353] env[61987]: DEBUG oslo_concurrency.lockutils [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "refresh_cache-056c181c-37da-49ac-ae1f-f7f2032c579a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.433876] env[61987]: DEBUG oslo_concurrency.lockutils [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired lock "refresh_cache-056c181c-37da-49ac-ae1f-f7f2032c579a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.434382] env[61987]: DEBUG nova.network.neutron [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 856.436387] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Acquiring lock "98d895ee-1d1f-49cd-a055-fe6c8462a692" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.436687] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Lock "98d895ee-1d1f-49cd-a055-fe6c8462a692" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.436961] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Acquiring lock "98d895ee-1d1f-49cd-a055-fe6c8462a692-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.437519] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Lock "98d895ee-1d1f-49cd-a055-fe6c8462a692-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.437519] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Lock "98d895ee-1d1f-49cd-a055-fe6c8462a692-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.440397] env[61987]: INFO nova.compute.manager [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Terminating instance [ 856.553343] env[61987]: DEBUG nova.compute.manager [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Received event network-changed-ebd9819d-a92d-4034-a00b-af37dbb338ab {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 856.556635] env[61987]: DEBUG nova.compute.manager [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Refreshing instance network info cache due to event network-changed-ebd9819d-a92d-4034-a00b-af37dbb338ab. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 856.556635] env[61987]: DEBUG oslo_concurrency.lockutils [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] Acquiring lock "refresh_cache-1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.556635] env[61987]: DEBUG oslo_concurrency.lockutils [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] Acquired lock "refresh_cache-1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.556635] env[61987]: DEBUG nova.network.neutron [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Refreshing network info cache for port ebd9819d-a92d-4034-a00b-af37dbb338ab {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 856.594717] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Releasing lock "refresh_cache-1bba4036-5211-4a43-b2e6-5276851335e2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.595675] env[61987]: DEBUG nova.compute.manager [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Instance network_info: |[{"id": "1460c9a3-914a-4c65-a2c3-601f1e07788e", "address": "fa:16:3e:c2:75:a9", "network": {"id": "8dfc473a-6d57-4169-9338-49a9565bb93d", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-422412073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c3d31c138df453486a2b1a6b70a84d4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1460c9a3-91", "ovs_interfaceid": "1460c9a3-914a-4c65-a2c3-601f1e07788e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 856.595817] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:75:a9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ae4e3171-21cd-4094-b6cf-81bf366c75bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1460c9a3-914a-4c65-a2c3-601f1e07788e', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 856.604255] env[61987]: DEBUG oslo.service.loopingcall [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 856.606300] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 856.610624] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-44367164-8071-4e1c-8c0d-0883064b78ae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.626802] env[61987]: DEBUG oslo_concurrency.lockutils [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] Releasing lock "refresh_cache-ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.627079] env[61987]: DEBUG nova.compute.manager [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Received event network-changed-69ce8bb3-dc28-422e-96a8-833062bd6b77 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 856.627268] env[61987]: DEBUG nova.compute.manager [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Refreshing instance network info cache due to event network-changed-69ce8bb3-dc28-422e-96a8-833062bd6b77. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 856.627489] env[61987]: DEBUG oslo_concurrency.lockutils [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] Acquiring lock "refresh_cache-98d895ee-1d1f-49cd-a055-fe6c8462a692" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.627643] env[61987]: DEBUG oslo_concurrency.lockutils [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] Acquired lock "refresh_cache-98d895ee-1d1f-49cd-a055-fe6c8462a692" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.627998] env[61987]: DEBUG nova.network.neutron [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Refreshing network info cache for port 69ce8bb3-dc28-422e-96a8-833062bd6b77 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 856.630768] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "interface-1f29d5c1-bd01-47dc-9dcf-4e89791120d0-ebd9819d-a92d-4034-a00b-af37dbb338ab" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.631059] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "interface-1f29d5c1-bd01-47dc-9dcf-4e89791120d0-ebd9819d-a92d-4034-a00b-af37dbb338ab" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.640394] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 856.640394] env[61987]: value = "task-1061804" [ 856.640394] env[61987]: _type = "Task" [ 856.640394] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.659535] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061804, 'name': CreateVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.760395] env[61987]: DEBUG oslo_vmware.api [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061802, 'name': PowerOffVM_Task, 'duration_secs': 0.261449} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.760747] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 856.764015] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 856.764015] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5b759bbb-febd-49bd-a4a4-04f4e1eeab1c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.906266] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbac0a58-9e50-40aa-9820-82e1340b6aa4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.922635] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c635b206-5893-4561-93b1-4b6f7f4771c9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.925932] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061801, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.573709} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.926128] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05/ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 856.927033] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 856.927217] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fb4ebfb0-313e-4743-b1c9-44b03f02ed95 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.960587] env[61987]: DEBUG nova.compute.manager [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 856.961034] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 856.964089] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f02d5119-1478-4a75-87b4-6b7c80a37da3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.968739] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e89fcbd9-4ab5-49db-8969-9365b47b8eb4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.974287] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 856.974287] env[61987]: value = "task-1061806" [ 856.974287] env[61987]: _type = "Task" [ 856.974287] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.988204] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 856.992472] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a68a7c5-ea8e-47d9-9d40-c3c37656ca2f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.999593] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc73de9a-b444-4d46-9520-b43d18bffd82 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.002252] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 857.002663] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 857.002974] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Deleting the datastore file [datastore1] 8f9aab3e-96ee-44c6-8d25-706a0ff857c4 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 857.007774] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2753590b-245b-4f22-978d-641260815e23 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.010507] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061806, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.026330] env[61987]: DEBUG nova.compute.provider_tree [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.032263] env[61987]: DEBUG oslo_vmware.api [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 857.032263] env[61987]: value = "task-1061808" [ 857.032263] env[61987]: _type = "Task" [ 857.032263] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.032768] env[61987]: DEBUG oslo_vmware.api [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Waiting for the task: (returnval){ [ 857.032768] env[61987]: value = "task-1061807" [ 857.032768] env[61987]: _type = "Task" [ 857.032768] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.054668] env[61987]: DEBUG oslo_vmware.api [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061807, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.055393] env[61987]: DEBUG oslo_vmware.api [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061808, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.139982] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.140192] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.141618] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aab9e4e8-12c3-46ae-8d8a-142624faa3b4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.154588] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061804, 'name': CreateVM_Task, 'duration_secs': 0.345991} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.167040] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 857.171615] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.171813] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.172203] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 857.173025] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89c1b205-52a8-4ff9-95ee-66fef08bd77f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.180912] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab6d0b07-f504-46bf-8240-a8e33dc2a88e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.183245] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "769ecb4d-8fdd-4993-8a29-84c30d1089b2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.183477] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "769ecb4d-8fdd-4993-8a29-84c30d1089b2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.189692] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 857.189692] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52ba7f1b-7948-a6ab-a5cb-5bc0ae93387b" [ 857.189692] env[61987]: _type = "Task" [ 857.189692] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.213159] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Reconfiguring VM to detach interface {{(pid=61987) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 857.221926] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b1b98b37-ee6e-4427-8f98-1219824f195e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.247424] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52ba7f1b-7948-a6ab-a5cb-5bc0ae93387b, 'name': SearchDatastore_Task, 'duration_secs': 0.012648} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.248745] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.249014] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 857.249289] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.249475] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.249634] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 857.249933] env[61987]: DEBUG oslo_vmware.api [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 857.249933] env[61987]: value = "task-1061809" [ 857.249933] env[61987]: _type = "Task" [ 857.249933] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.250233] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0e7318fb-249f-47e5-95bd-4a780a0f1bfa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.263571] env[61987]: DEBUG oslo_vmware.api [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061809, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.266086] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 857.266322] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 857.267202] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-884e1e54-1af8-42c8-84bb-d077a06e9974 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.274274] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 857.274274] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]525d894b-8469-377a-29f1-f82c84d8aeac" [ 857.274274] env[61987]: _type = "Task" [ 857.274274] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.284264] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]525d894b-8469-377a-29f1-f82c84d8aeac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.425793] env[61987]: DEBUG nova.network.neutron [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Updated VIF entry in instance network info cache for port 69ce8bb3-dc28-422e-96a8-833062bd6b77. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 857.425793] env[61987]: DEBUG nova.network.neutron [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Updating instance_info_cache with network_info: [{"id": "69ce8bb3-dc28-422e-96a8-833062bd6b77", "address": "fa:16:3e:ab:10:c2", "network": {"id": "35162a8d-60d3-465b-9330-64bbd235d6a2", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-981766067-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8dffc4d03e514fb8932b80718375084b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68add7d6-c025-46fa-84d3-9c589adb63e4", "external-id": "nsx-vlan-transportzone-961", "segmentation_id": 961, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69ce8bb3-dc", "ovs_interfaceid": "69ce8bb3-dc28-422e-96a8-833062bd6b77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.487041] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061806, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075349} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.488717] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 857.489638] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42bca62-1bd3-4030-b93b-19a9375dd337 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.514692] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05/ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 857.515083] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b78def12-e23b-4dc9-b796-430aa4b6c7e3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.530246] env[61987]: DEBUG nova.scheduler.client.report [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 857.544476] env[61987]: DEBUG nova.network.neutron [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Updating instance_info_cache with network_info: [{"id": "e88ac5c4-67fa-4013-ac31-ca59251955b5", "address": "fa:16:3e:14:12:b9", "network": {"id": "ca68bf96-3f18-45ee-80fb-64a1260ed996", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-733595377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd47eb44bc334bf3ae5813905903ecbc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a27fd90b-16a5-43af-bede-ae36762ece00", "external-id": "nsx-vlan-transportzone-197", "segmentation_id": 197, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape88ac5c4-67", "ovs_interfaceid": "e88ac5c4-67fa-4013-ac31-ca59251955b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.547850] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 857.547850] env[61987]: value = "task-1061810" [ 857.547850] env[61987]: _type = "Task" [ 857.547850] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.560119] env[61987]: DEBUG oslo_vmware.api [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061808, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.203183} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.560480] env[61987]: DEBUG oslo_vmware.api [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061807, 'name': PowerOffVM_Task, 'duration_secs': 0.292901} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.562612] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 857.562861] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 857.563091] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 857.563336] env[61987]: INFO nova.compute.manager [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Took 1.34 seconds to destroy the instance on the hypervisor. [ 857.563624] env[61987]: DEBUG oslo.service.loopingcall [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 857.563873] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 857.564479] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 857.568282] env[61987]: DEBUG nova.compute.manager [-] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 857.568573] env[61987]: DEBUG nova.network.neutron [-] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 857.572705] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d2d573e4-9361-482b-92c9-5279e5739bc1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.574638] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061810, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.630019] env[61987]: DEBUG nova.network.neutron [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Updated VIF entry in instance network info cache for port ebd9819d-a92d-4034-a00b-af37dbb338ab. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 857.630556] env[61987]: DEBUG nova.network.neutron [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Updating instance_info_cache with network_info: [{"id": "e7d811c0-d77e-448d-86fe-acdb61c04284", "address": "fa:16:3e:b0:6d:8c", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7d811c0-d7", "ovs_interfaceid": "e7d811c0-d77e-448d-86fe-acdb61c04284", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "ebd9819d-a92d-4034-a00b-af37dbb338ab", "address": "fa:16:3e:36:15:13", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapebd9819d-a9", "ovs_interfaceid": "ebd9819d-a92d-4034-a00b-af37dbb338ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.643987] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 857.644277] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 857.645490] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Deleting the datastore file [datastore2] 98d895ee-1d1f-49cd-a055-fe6c8462a692 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 857.645580] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-207d8054-5e94-467f-9275-0504377774ae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.657042] env[61987]: DEBUG oslo_vmware.api [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Waiting for the task: (returnval){ [ 857.657042] env[61987]: value = "task-1061812" [ 857.657042] env[61987]: _type = "Task" [ 857.657042] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.666891] env[61987]: DEBUG oslo_vmware.api [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061812, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.685845] env[61987]: DEBUG nova.compute.manager [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 857.764153] env[61987]: DEBUG oslo_vmware.api [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061809, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.787271] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]525d894b-8469-377a-29f1-f82c84d8aeac, 'name': SearchDatastore_Task, 'duration_secs': 0.016612} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.788968] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ca5fd0c-f213-4fae-9f45-c008033c1880 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.794410] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 857.794410] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]523d89ef-0cdd-fdd7-2c93-1e0a36e35264" [ 857.794410] env[61987]: _type = "Task" [ 857.794410] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.803865] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523d89ef-0cdd-fdd7-2c93-1e0a36e35264, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.927069] env[61987]: DEBUG oslo_concurrency.lockutils [req-25a8f356-bbe8-4faf-9e18-154a3f8be42a req-067ce191-5ad4-4847-9f1e-6f760b5533fb service nova] Releasing lock "refresh_cache-98d895ee-1d1f-49cd-a055-fe6c8462a692" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.035713] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.399s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.036484] env[61987]: DEBUG nova.compute.manager [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 858.047756] env[61987]: DEBUG oslo_concurrency.lockutils [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Releasing lock "refresh_cache-056c181c-37da-49ac-ae1f-f7f2032c579a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.060563] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061810, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.133843] env[61987]: DEBUG oslo_concurrency.lockutils [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] Releasing lock "refresh_cache-1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.134294] env[61987]: DEBUG nova.compute.manager [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Received event network-vif-plugged-1460c9a3-914a-4c65-a2c3-601f1e07788e {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 858.134485] env[61987]: DEBUG oslo_concurrency.lockutils [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] Acquiring lock "1bba4036-5211-4a43-b2e6-5276851335e2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.134731] env[61987]: DEBUG oslo_concurrency.lockutils [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] Lock "1bba4036-5211-4a43-b2e6-5276851335e2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.134924] env[61987]: DEBUG oslo_concurrency.lockutils [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] Lock "1bba4036-5211-4a43-b2e6-5276851335e2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.135130] env[61987]: DEBUG nova.compute.manager [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] No waiting events found dispatching network-vif-plugged-1460c9a3-914a-4c65-a2c3-601f1e07788e {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 858.135404] env[61987]: WARNING nova.compute.manager [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Received unexpected event network-vif-plugged-1460c9a3-914a-4c65-a2c3-601f1e07788e for instance with vm_state building and task_state spawning. [ 858.135613] env[61987]: DEBUG nova.compute.manager [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Received event network-changed-1460c9a3-914a-4c65-a2c3-601f1e07788e {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 858.135799] env[61987]: DEBUG nova.compute.manager [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Refreshing instance network info cache due to event network-changed-1460c9a3-914a-4c65-a2c3-601f1e07788e. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 858.135998] env[61987]: DEBUG oslo_concurrency.lockutils [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] Acquiring lock "refresh_cache-1bba4036-5211-4a43-b2e6-5276851335e2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.136451] env[61987]: DEBUG oslo_concurrency.lockutils [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] Acquired lock "refresh_cache-1bba4036-5211-4a43-b2e6-5276851335e2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.136451] env[61987]: DEBUG nova.network.neutron [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Refreshing network info cache for port 1460c9a3-914a-4c65-a2c3-601f1e07788e {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 858.166925] env[61987]: DEBUG oslo_vmware.api [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Task: {'id': task-1061812, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.369183} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.166925] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 858.166925] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 858.166925] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 858.166925] env[61987]: INFO nova.compute.manager [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Took 1.21 seconds to destroy the instance on the hypervisor. [ 858.167215] env[61987]: DEBUG oslo.service.loopingcall [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.167271] env[61987]: DEBUG nova.compute.manager [-] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 858.167321] env[61987]: DEBUG nova.network.neutron [-] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 858.208731] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.209122] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.211285] env[61987]: INFO nova.compute.claims [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 858.266293] env[61987]: DEBUG oslo_vmware.api [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061809, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.304966] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523d89ef-0cdd-fdd7-2c93-1e0a36e35264, 'name': SearchDatastore_Task, 'duration_secs': 0.014362} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.305404] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.305601] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 1bba4036-5211-4a43-b2e6-5276851335e2/1bba4036-5211-4a43-b2e6-5276851335e2.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 858.305851] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7936dbe0-9370-452c-847d-bacaa2063170 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.313258] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 858.313258] env[61987]: value = "task-1061813" [ 858.313258] env[61987]: _type = "Task" [ 858.313258] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.321767] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061813, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.542789] env[61987]: DEBUG nova.compute.utils [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 858.544124] env[61987]: DEBUG nova.compute.manager [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 858.544337] env[61987]: DEBUG nova.network.neutron [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 858.550775] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 858.552453] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c987c0-66fc-48f0-9b01-14ba76b27107 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.563764] env[61987]: DEBUG nova.compute.manager [req-971d2438-499e-46b4-8a18-97fc11894e8c req-4897590a-b625-48eb-9ca9-c6dc1c8400a1 service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Received event network-vif-deleted-69ce8bb3-dc28-422e-96a8-833062bd6b77 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 858.563995] env[61987]: INFO nova.compute.manager [req-971d2438-499e-46b4-8a18-97fc11894e8c req-4897590a-b625-48eb-9ca9-c6dc1c8400a1 service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Neutron deleted interface 69ce8bb3-dc28-422e-96a8-833062bd6b77; detaching it from the instance and deleting it from the info cache [ 858.564378] env[61987]: DEBUG nova.network.neutron [req-971d2438-499e-46b4-8a18-97fc11894e8c req-4897590a-b625-48eb-9ca9-c6dc1c8400a1 service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.572789] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 858.579389] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-05d46a02-bcfd-4378-a822-45aa86be121f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.582599] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061810, 'name': ReconfigVM_Task, 'duration_secs': 0.577392} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.583132] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Reconfigured VM instance instance-0000004b to attach disk [datastore2] ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05/ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 858.585819] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-31c60007-bb10-4c4b-8104-ae852972e291 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.596127] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 858.596127] env[61987]: value = "task-1061815" [ 858.596127] env[61987]: _type = "Task" [ 858.596127] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.603977] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061815, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.607683] env[61987]: DEBUG nova.compute.manager [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Received event network-changed-0a4d6904-241e-451e-86d4-2cec32cf3b25 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 858.607981] env[61987]: DEBUG nova.compute.manager [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Refreshing instance network info cache due to event network-changed-0a4d6904-241e-451e-86d4-2cec32cf3b25. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 858.608229] env[61987]: DEBUG oslo_concurrency.lockutils [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] Acquiring lock "refresh_cache-742297eb-c45d-4f45-85d2-e9bdd3106d1e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.608384] env[61987]: DEBUG oslo_concurrency.lockutils [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] Acquired lock "refresh_cache-742297eb-c45d-4f45-85d2-e9bdd3106d1e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.608550] env[61987]: DEBUG nova.network.neutron [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Refreshing network info cache for port 0a4d6904-241e-451e-86d4-2cec32cf3b25 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 858.654090] env[61987]: DEBUG nova.policy [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '102f79eb06304aebbe40ff6ae9f89376', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e43b1d992c07429685f7c89f1a8cfc9a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 858.660370] env[61987]: DEBUG oslo_vmware.rw_handles [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52485f60-c3dd-e925-b0d2-ce13f8afb9c7/disk-0.vmdk. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 858.661841] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d7b4adb-7a45-4585-b46b-4f492aea1934 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.669571] env[61987]: DEBUG oslo_vmware.rw_handles [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52485f60-c3dd-e925-b0d2-ce13f8afb9c7/disk-0.vmdk is in state: ready. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 858.669901] env[61987]: ERROR oslo_vmware.rw_handles [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52485f60-c3dd-e925-b0d2-ce13f8afb9c7/disk-0.vmdk due to incomplete transfer. [ 858.670229] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-65371dec-504c-49dd-898b-83ee51ae2202 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.684471] env[61987]: DEBUG oslo_vmware.rw_handles [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52485f60-c3dd-e925-b0d2-ce13f8afb9c7/disk-0.vmdk. {{(pid=61987) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 858.684861] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Uploaded image 87a46110-ae87-47e9-975e-48f434a17e9a to the Glance image server {{(pid=61987) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 858.689274] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Destroying the VM {{(pid=61987) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 858.691021] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-90d15d47-3084-4d7d-8a9b-067756c273c6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.697844] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 858.698173] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 858.698403] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Deleting the datastore file [datastore2] 056c181c-37da-49ac-ae1f-f7f2032c579a {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 858.701271] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ca3acad5-6518-4be4-a169-b7ffbd59a72b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.710111] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 858.710111] env[61987]: value = "task-1061816" [ 858.710111] env[61987]: _type = "Task" [ 858.710111] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.712432] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 858.712432] env[61987]: value = "task-1061817" [ 858.712432] env[61987]: _type = "Task" [ 858.712432] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.732404] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061817, 'name': Destroy_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.740208] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061816, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.770347] env[61987]: DEBUG oslo_vmware.api [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061809, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.824429] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061813, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.866512] env[61987]: DEBUG nova.network.neutron [-] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.952865] env[61987]: DEBUG nova.network.neutron [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Updated VIF entry in instance network info cache for port 1460c9a3-914a-4c65-a2c3-601f1e07788e. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 858.953329] env[61987]: DEBUG nova.network.neutron [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Updating instance_info_cache with network_info: [{"id": "1460c9a3-914a-4c65-a2c3-601f1e07788e", "address": "fa:16:3e:c2:75:a9", "network": {"id": "8dfc473a-6d57-4169-9338-49a9565bb93d", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-422412073-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0c3d31c138df453486a2b1a6b70a84d4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ae4e3171-21cd-4094-b6cf-81bf366c75bd", "external-id": "nsx-vlan-transportzone-193", "segmentation_id": 193, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1460c9a3-91", "ovs_interfaceid": "1460c9a3-914a-4c65-a2c3-601f1e07788e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.020029] env[61987]: DEBUG nova.network.neutron [-] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.046997] env[61987]: DEBUG nova.compute.manager [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 859.066259] env[61987]: DEBUG nova.network.neutron [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Successfully created port: e4942ff0-6c8c-4c29-850d-778e6caa8e44 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 859.068996] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0bda533b-9eb7-4dc6-bbfa-24b584e43c50 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.078632] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b838089b-e134-41e3-a6fa-7afacc420f3e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.104478] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061815, 'name': Rename_Task, 'duration_secs': 0.174053} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.114476] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 859.117642] env[61987]: DEBUG nova.compute.manager [req-971d2438-499e-46b4-8a18-97fc11894e8c req-4897590a-b625-48eb-9ca9-c6dc1c8400a1 service nova] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Detach interface failed, port_id=69ce8bb3-dc28-422e-96a8-833062bd6b77, reason: Instance 98d895ee-1d1f-49cd-a055-fe6c8462a692 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 859.117642] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0ea788b1-b247-4aa6-8111-96d1f2779607 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.124285] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 859.124285] env[61987]: value = "task-1061818" [ 859.124285] env[61987]: _type = "Task" [ 859.124285] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.133713] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061818, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.224302] env[61987]: DEBUG oslo_vmware.api [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061816, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.315009} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.224918] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 859.225510] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 859.225727] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 859.233259] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061817, 'name': Destroy_Task, 'duration_secs': 0.484134} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.233965] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Destroyed the VM [ 859.234250] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Deleting Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 859.234664] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-f7de848f-e96c-41e4-ad91-fe8d84de108c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.240021] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 859.240021] env[61987]: value = "task-1061819" [ 859.240021] env[61987]: _type = "Task" [ 859.240021] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.250793] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061819, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.262275] env[61987]: DEBUG oslo_vmware.api [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061809, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.265667] env[61987]: INFO nova.scheduler.client.report [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Deleted allocations for instance 056c181c-37da-49ac-ae1f-f7f2032c579a [ 859.331013] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061813, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.570042} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.331312] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 1bba4036-5211-4a43-b2e6-5276851335e2/1bba4036-5211-4a43-b2e6-5276851335e2.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 859.331537] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 859.331799] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-62e5afb0-3a1e-4e9b-b89f-e8fde6609c63 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.341581] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 859.341581] env[61987]: value = "task-1061820" [ 859.341581] env[61987]: _type = "Task" [ 859.341581] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.351584] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061820, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.371463] env[61987]: INFO nova.compute.manager [-] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Took 1.80 seconds to deallocate network for instance. [ 859.443289] env[61987]: DEBUG nova.network.neutron [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Updated VIF entry in instance network info cache for port 0a4d6904-241e-451e-86d4-2cec32cf3b25. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 859.443683] env[61987]: DEBUG nova.network.neutron [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Updating instance_info_cache with network_info: [{"id": "0a4d6904-241e-451e-86d4-2cec32cf3b25", "address": "fa:16:3e:bb:78:bb", "network": {"id": "26d453d1-8edf-45e1-9e49-2a3225de6633", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1400564876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.132", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b77fbbfc7f6437083f9f30e8fdeeb35", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0a4d6904-24", "ovs_interfaceid": "0a4d6904-241e-451e-86d4-2cec32cf3b25", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.458838] env[61987]: DEBUG oslo_concurrency.lockutils [req-26ae9634-5d69-4c05-bb0c-197a41943a81 req-ba035244-5cd6-48ab-b28f-e4675787169c service nova] Releasing lock "refresh_cache-1bba4036-5211-4a43-b2e6-5276851335e2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.517779] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35dcf970-59c6-412e-a41d-aaeee676afe6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.522217] env[61987]: INFO nova.compute.manager [-] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Took 1.35 seconds to deallocate network for instance. [ 859.529694] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26452feb-9fc5-4b50-99c7-c76d25a95ccd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.569799] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2d3fda9-8f76-4cc8-a3ec-f6c27c4eef04 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.578975] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a48123-bec0-4865-a64b-b6b56aae4b43 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.596132] env[61987]: DEBUG nova.compute.provider_tree [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.635920] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061818, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.750575] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061819, 'name': RemoveSnapshot_Task, 'duration_secs': 0.409021} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.750680] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Deleted Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 859.750961] env[61987]: DEBUG nova.compute.manager [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 859.751778] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71169c7a-8a9f-486d-bc1e-a6bc35ebf6ae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.767640] env[61987]: DEBUG oslo_vmware.api [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061809, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.770890] env[61987]: DEBUG oslo_concurrency.lockutils [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.851224] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061820, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067547} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.852210] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 859.853154] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-405592a0-caf4-4f3e-85f0-6c847611222b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.878456] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 1bba4036-5211-4a43-b2e6-5276851335e2/1bba4036-5211-4a43-b2e6-5276851335e2.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 859.879670] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.880094] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-769cce8b-49da-4fd6-b846-96e934da154f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.904143] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 859.904143] env[61987]: value = "task-1061821" [ 859.904143] env[61987]: _type = "Task" [ 859.904143] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.915767] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061821, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.949869] env[61987]: DEBUG oslo_concurrency.lockutils [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] Releasing lock "refresh_cache-742297eb-c45d-4f45-85d2-e9bdd3106d1e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.949869] env[61987]: DEBUG nova.compute.manager [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Received event network-vif-deleted-ac11f37b-72cc-4fa6-a61a-4535486b9d7f {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 859.949869] env[61987]: INFO nova.compute.manager [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Neutron deleted interface ac11f37b-72cc-4fa6-a61a-4535486b9d7f; detaching it from the instance and deleting it from the info cache [ 859.949869] env[61987]: DEBUG nova.network.neutron [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.036438] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.074563] env[61987]: DEBUG nova.compute.manager [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 860.103110] env[61987]: DEBUG nova.scheduler.client.report [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 860.122109] env[61987]: DEBUG nova.virt.hardware [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 860.122109] env[61987]: DEBUG nova.virt.hardware [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 860.122109] env[61987]: DEBUG nova.virt.hardware [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 860.122109] env[61987]: DEBUG nova.virt.hardware [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 860.122612] env[61987]: DEBUG nova.virt.hardware [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 860.122962] env[61987]: DEBUG nova.virt.hardware [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 860.123309] env[61987]: DEBUG nova.virt.hardware [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 860.123606] env[61987]: DEBUG nova.virt.hardware [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 860.126380] env[61987]: DEBUG nova.virt.hardware [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 860.126380] env[61987]: DEBUG nova.virt.hardware [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 860.126380] env[61987]: DEBUG nova.virt.hardware [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 860.126380] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0baaa581-60d4-43fc-9a01-7ca215f198ab {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.143020] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b2c10e7-fdee-4ef3-a772-3ca2a5fa65c8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.147503] env[61987]: DEBUG oslo_vmware.api [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061818, 'name': PowerOnVM_Task, 'duration_secs': 0.518024} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.148165] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 860.148493] env[61987]: INFO nova.compute.manager [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Took 10.20 seconds to spawn the instance on the hypervisor. [ 860.148786] env[61987]: DEBUG nova.compute.manager [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 860.149915] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc8f7d9-446f-4501-a412-e9cd313b8230 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.266921] env[61987]: DEBUG oslo_vmware.api [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061809, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.267395] env[61987]: INFO nova.compute.manager [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Shelve offloading [ 860.414999] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061821, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.454901] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e1f739d1-c118-42cb-a4d8-2aa3ae5e3543 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.463803] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e7e2fcc-a78a-43c3-b57d-6e840446e2b4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.494952] env[61987]: DEBUG nova.compute.manager [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Detach interface failed, port_id=ac11f37b-72cc-4fa6-a61a-4535486b9d7f, reason: Instance 8f9aab3e-96ee-44c6-8d25-706a0ff857c4 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 860.495267] env[61987]: DEBUG nova.compute.manager [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Received event network-vif-unplugged-e88ac5c4-67fa-4013-ac31-ca59251955b5 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 860.495541] env[61987]: DEBUG oslo_concurrency.lockutils [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] Acquiring lock "056c181c-37da-49ac-ae1f-f7f2032c579a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.495760] env[61987]: DEBUG oslo_concurrency.lockutils [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] Lock "056c181c-37da-49ac-ae1f-f7f2032c579a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.495946] env[61987]: DEBUG oslo_concurrency.lockutils [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] Lock "056c181c-37da-49ac-ae1f-f7f2032c579a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.496143] env[61987]: DEBUG nova.compute.manager [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] No waiting events found dispatching network-vif-unplugged-e88ac5c4-67fa-4013-ac31-ca59251955b5 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 860.496337] env[61987]: WARNING nova.compute.manager [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Received unexpected event network-vif-unplugged-e88ac5c4-67fa-4013-ac31-ca59251955b5 for instance with vm_state shelved and task_state shelving_offloading. [ 860.496513] env[61987]: DEBUG nova.compute.manager [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Received event network-changed-e88ac5c4-67fa-4013-ac31-ca59251955b5 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 860.496673] env[61987]: DEBUG nova.compute.manager [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Refreshing instance network info cache due to event network-changed-e88ac5c4-67fa-4013-ac31-ca59251955b5. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 860.496864] env[61987]: DEBUG oslo_concurrency.lockutils [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] Acquiring lock "refresh_cache-056c181c-37da-49ac-ae1f-f7f2032c579a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.497062] env[61987]: DEBUG oslo_concurrency.lockutils [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] Acquired lock "refresh_cache-056c181c-37da-49ac-ae1f-f7f2032c579a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.497176] env[61987]: DEBUG nova.network.neutron [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Refreshing network info cache for port e88ac5c4-67fa-4013-ac31-ca59251955b5 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 860.616227] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.407s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.616801] env[61987]: DEBUG nova.compute.manager [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 860.620305] env[61987]: DEBUG oslo_concurrency.lockutils [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.849s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.620588] env[61987]: DEBUG nova.objects.instance [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lazy-loading 'resources' on Instance uuid 056c181c-37da-49ac-ae1f-f7f2032c579a {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 860.680481] env[61987]: INFO nova.compute.manager [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Took 22.17 seconds to build instance. [ 860.767177] env[61987]: DEBUG oslo_vmware.api [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061809, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.768419] env[61987]: DEBUG nova.compute.manager [req-956aedd6-c09b-470c-bb13-7e127464033a req-27a8d06c-5759-43ab-bea2-9fb845f23dd2 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Received event network-vif-plugged-e4942ff0-6c8c-4c29-850d-778e6caa8e44 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 860.768641] env[61987]: DEBUG oslo_concurrency.lockutils [req-956aedd6-c09b-470c-bb13-7e127464033a req-27a8d06c-5759-43ab-bea2-9fb845f23dd2 service nova] Acquiring lock "455562c7-9fc1-4625-95b0-5eb8c9c8ccaa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.769038] env[61987]: DEBUG oslo_concurrency.lockutils [req-956aedd6-c09b-470c-bb13-7e127464033a req-27a8d06c-5759-43ab-bea2-9fb845f23dd2 service nova] Lock "455562c7-9fc1-4625-95b0-5eb8c9c8ccaa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.769260] env[61987]: DEBUG oslo_concurrency.lockutils [req-956aedd6-c09b-470c-bb13-7e127464033a req-27a8d06c-5759-43ab-bea2-9fb845f23dd2 service nova] Lock "455562c7-9fc1-4625-95b0-5eb8c9c8ccaa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.769441] env[61987]: DEBUG nova.compute.manager [req-956aedd6-c09b-470c-bb13-7e127464033a req-27a8d06c-5759-43ab-bea2-9fb845f23dd2 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] No waiting events found dispatching network-vif-plugged-e4942ff0-6c8c-4c29-850d-778e6caa8e44 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 860.769611] env[61987]: WARNING nova.compute.manager [req-956aedd6-c09b-470c-bb13-7e127464033a req-27a8d06c-5759-43ab-bea2-9fb845f23dd2 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Received unexpected event network-vif-plugged-e4942ff0-6c8c-4c29-850d-778e6caa8e44 for instance with vm_state building and task_state spawning. [ 860.770396] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 860.770652] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f81b556-86f6-45c9-ba86-cabd8ee5dd93 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.778225] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 860.778225] env[61987]: value = "task-1061822" [ 860.778225] env[61987]: _type = "Task" [ 860.778225] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.788105] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] VM already powered off {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 860.788621] env[61987]: DEBUG nova.compute.manager [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 860.789416] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da69a17-3767-419e-855f-f7b825674edd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.796202] env[61987]: DEBUG oslo_concurrency.lockutils [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "refresh_cache-51a36102-795e-47b7-b96a-857e54dc703e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.796379] env[61987]: DEBUG oslo_concurrency.lockutils [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquired lock "refresh_cache-51a36102-795e-47b7-b96a-857e54dc703e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.796557] env[61987]: DEBUG nova.network.neutron [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 860.924017] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061821, 'name': ReconfigVM_Task, 'duration_secs': 0.856569} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.924017] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 1bba4036-5211-4a43-b2e6-5276851335e2/1bba4036-5211-4a43-b2e6-5276851335e2.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 860.924017] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ec9c4306-b0c3-4843-98b7-aa903c0f7b3a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.930056] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 860.930056] env[61987]: value = "task-1061823" [ 860.930056] env[61987]: _type = "Task" [ 860.930056] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.945191] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061823, 'name': Rename_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.124336] env[61987]: DEBUG nova.compute.utils [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 861.125998] env[61987]: DEBUG nova.objects.instance [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lazy-loading 'numa_topology' on Instance uuid 056c181c-37da-49ac-ae1f-f7f2032c579a {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 861.127283] env[61987]: DEBUG nova.compute.manager [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 861.127458] env[61987]: DEBUG nova.network.neutron [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 861.181209] env[61987]: DEBUG oslo_concurrency.lockutils [None req-69648c74-713e-4cc9-a983-8d8d157c9091 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.687s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.204237] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26676183-2273-4b3f-8234-5205ed2e6e47 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "056c181c-37da-49ac-ae1f-f7f2032c579a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.224387] env[61987]: DEBUG nova.policy [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e16f557dfdb04406b3fc772b891f05a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '22e62741abfb4471bbc7e11ae7d407ca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 861.265975] env[61987]: DEBUG oslo_vmware.api [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061809, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.428253] env[61987]: DEBUG nova.network.neutron [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Updated VIF entry in instance network info cache for port e88ac5c4-67fa-4013-ac31-ca59251955b5. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 861.428649] env[61987]: DEBUG nova.network.neutron [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Updating instance_info_cache with network_info: [{"id": "e88ac5c4-67fa-4013-ac31-ca59251955b5", "address": "fa:16:3e:14:12:b9", "network": {"id": "ca68bf96-3f18-45ee-80fb-64a1260ed996", "bridge": null, "label": "tempest-DeleteServersTestJSON-733595377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd47eb44bc334bf3ae5813905903ecbc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tape88ac5c4-67", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.442505] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061823, 'name': Rename_Task, 'duration_secs': 0.143102} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.443657] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 861.443657] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1523388f-1088-47f6-954e-88386969211e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.450868] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 861.450868] env[61987]: value = "task-1061824" [ 861.450868] env[61987]: _type = "Task" [ 861.450868] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.461560] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061824, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.488020] env[61987]: DEBUG nova.network.neutron [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Successfully updated port: e4942ff0-6c8c-4c29-850d-778e6caa8e44 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 861.628418] env[61987]: DEBUG nova.compute.manager [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 861.632488] env[61987]: DEBUG nova.objects.base [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Object Instance<056c181c-37da-49ac-ae1f-f7f2032c579a> lazy-loaded attributes: resources,numa_topology {{(pid=61987) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 861.766069] env[61987]: DEBUG oslo_vmware.api [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061809, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.814671] env[61987]: DEBUG nova.compute.manager [req-babb27e3-8dcb-4aeb-83f2-482d19036a1e req-938761b4-329c-4cf7-a3af-671b656d39a6 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Received event network-changed-e4942ff0-6c8c-4c29-850d-778e6caa8e44 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 861.815355] env[61987]: DEBUG nova.compute.manager [req-babb27e3-8dcb-4aeb-83f2-482d19036a1e req-938761b4-329c-4cf7-a3af-671b656d39a6 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Refreshing instance network info cache due to event network-changed-e4942ff0-6c8c-4c29-850d-778e6caa8e44. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 861.815355] env[61987]: DEBUG oslo_concurrency.lockutils [req-babb27e3-8dcb-4aeb-83f2-482d19036a1e req-938761b4-329c-4cf7-a3af-671b656d39a6 service nova] Acquiring lock "refresh_cache-455562c7-9fc1-4625-95b0-5eb8c9c8ccaa" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.815355] env[61987]: DEBUG oslo_concurrency.lockutils [req-babb27e3-8dcb-4aeb-83f2-482d19036a1e req-938761b4-329c-4cf7-a3af-671b656d39a6 service nova] Acquired lock "refresh_cache-455562c7-9fc1-4625-95b0-5eb8c9c8ccaa" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.815631] env[61987]: DEBUG nova.network.neutron [req-babb27e3-8dcb-4aeb-83f2-482d19036a1e req-938761b4-329c-4cf7-a3af-671b656d39a6 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Refreshing network info cache for port e4942ff0-6c8c-4c29-850d-778e6caa8e44 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 861.853383] env[61987]: DEBUG nova.network.neutron [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Updating instance_info_cache with network_info: [{"id": "4e18ece6-5812-4c5c-827d-b3ee7d399008", "address": "fa:16:3e:1a:e6:c9", "network": {"id": "abcd2857-4c5a-4546-a003-858ddb17b0e6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-124838733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3fd98395e57149c987f82585296751b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e18ece6-58", "ovs_interfaceid": "4e18ece6-5812-4c5c-827d-b3ee7d399008", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.892539] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b913822b-bb68-436b-b085-2d7ce20422ed {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.896593] env[61987]: DEBUG nova.network.neutron [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Successfully created port: a89e61c6-e262-4b30-a88c-a8731128abb1 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 861.902939] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76094cc7-1596-474c-b87d-4dcb8c21a757 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.933340] env[61987]: DEBUG oslo_concurrency.lockutils [req-cfd21ef2-2d30-4ffa-a523-79cc3281a1b2 req-30f5f812-b322-4eac-b8e5-2d89604db4bc service nova] Releasing lock "refresh_cache-056c181c-37da-49ac-ae1f-f7f2032c579a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.934547] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a9fa20b-e52e-4aa8-8bb2-8bcf1e8399de {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.942413] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14ff3936-445f-48c8-8a24-c26d3ac4e3bc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.958088] env[61987]: DEBUG nova.compute.provider_tree [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.967071] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061824, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.989893] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquiring lock "refresh_cache-455562c7-9fc1-4625-95b0-5eb8c9c8ccaa" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.268293] env[61987]: DEBUG oslo_vmware.api [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061809, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.351565] env[61987]: DEBUG nova.network.neutron [req-babb27e3-8dcb-4aeb-83f2-482d19036a1e req-938761b4-329c-4cf7-a3af-671b656d39a6 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 862.357286] env[61987]: DEBUG oslo_concurrency.lockutils [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Releasing lock "refresh_cache-51a36102-795e-47b7-b96a-857e54dc703e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.466095] env[61987]: DEBUG nova.scheduler.client.report [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 862.470897] env[61987]: DEBUG oslo_vmware.api [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061824, 'name': PowerOnVM_Task, 'duration_secs': 0.850911} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.471411] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 862.472016] env[61987]: INFO nova.compute.manager [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Took 8.23 seconds to spawn the instance on the hypervisor. [ 862.472016] env[61987]: DEBUG nova.compute.manager [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 862.472851] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-676eff63-14e3-42b9-925b-e4ab7439daea {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.494102] env[61987]: DEBUG nova.network.neutron [req-babb27e3-8dcb-4aeb-83f2-482d19036a1e req-938761b4-329c-4cf7-a3af-671b656d39a6 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.643302] env[61987]: DEBUG nova.compute.manager [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 862.678604] env[61987]: DEBUG nova.virt.hardware [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 862.678860] env[61987]: DEBUG nova.virt.hardware [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 862.679035] env[61987]: DEBUG nova.virt.hardware [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 862.679233] env[61987]: DEBUG nova.virt.hardware [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 862.679391] env[61987]: DEBUG nova.virt.hardware [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 862.679545] env[61987]: DEBUG nova.virt.hardware [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 862.679769] env[61987]: DEBUG nova.virt.hardware [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 862.679942] env[61987]: DEBUG nova.virt.hardware [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 862.680201] env[61987]: DEBUG nova.virt.hardware [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 862.680380] env[61987]: DEBUG nova.virt.hardware [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 862.680571] env[61987]: DEBUG nova.virt.hardware [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 862.681426] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80cee895-6a42-47ad-a337-b303e2e33a86 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.690328] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ecd6aa-5b1c-45c5-b86e-b4274ecd5202 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.744125] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 862.744125] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7444578-5aaf-4593-8d7f-a7aa8ad948e9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.751172] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 862.751625] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1ea3c92-9dda-4b26-a9aa-8bef062fe300 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.766763] env[61987]: DEBUG oslo_vmware.api [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061809, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.832928] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 862.833254] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 862.833495] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Deleting the datastore file [datastore2] 51a36102-795e-47b7-b96a-857e54dc703e {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 862.833893] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-61d32d72-859f-4885-89af-e33f74313fcc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.840217] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 862.840217] env[61987]: value = "task-1061826" [ 862.840217] env[61987]: _type = "Task" [ 862.840217] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.849726] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061826, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.939727] env[61987]: DEBUG nova.compute.manager [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Stashing vm_state: active {{(pid=61987) _prep_resize /opt/stack/nova/nova/compute/manager.py:6090}} [ 862.973342] env[61987]: DEBUG oslo_concurrency.lockutils [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.353s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.976151] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.096s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.976414] env[61987]: DEBUG nova.objects.instance [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lazy-loading 'resources' on Instance uuid 8f9aab3e-96ee-44c6-8d25-706a0ff857c4 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.994978] env[61987]: INFO nova.compute.manager [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Took 21.43 seconds to build instance. [ 862.996609] env[61987]: DEBUG oslo_concurrency.lockutils [req-babb27e3-8dcb-4aeb-83f2-482d19036a1e req-938761b4-329c-4cf7-a3af-671b656d39a6 service nova] Releasing lock "refresh_cache-455562c7-9fc1-4625-95b0-5eb8c9c8ccaa" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.997407] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquired lock "refresh_cache-455562c7-9fc1-4625-95b0-5eb8c9c8ccaa" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.997596] env[61987]: DEBUG nova.network.neutron [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 863.266965] env[61987]: DEBUG oslo_vmware.api [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061809, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.350381] env[61987]: DEBUG oslo_vmware.api [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061826, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13936} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.350747] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 863.350918] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 863.351078] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 863.381102] env[61987]: INFO nova.scheduler.client.report [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Deleted allocations for instance 51a36102-795e-47b7-b96a-857e54dc703e [ 863.462430] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.486438] env[61987]: DEBUG oslo_concurrency.lockutils [None req-972d91da-ebdd-445e-93ce-adc2cff4cb1b tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "056c181c-37da-49ac-ae1f-f7f2032c579a" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 26.355s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.487355] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26676183-2273-4b3f-8234-5205ed2e6e47 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "056c181c-37da-49ac-ae1f-f7f2032c579a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 2.283s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.487598] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26676183-2273-4b3f-8234-5205ed2e6e47 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "056c181c-37da-49ac-ae1f-f7f2032c579a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.487852] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26676183-2273-4b3f-8234-5205ed2e6e47 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "056c181c-37da-49ac-ae1f-f7f2032c579a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.488069] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26676183-2273-4b3f-8234-5205ed2e6e47 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "056c181c-37da-49ac-ae1f-f7f2032c579a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.492736] env[61987]: INFO nova.compute.manager [None req-26676183-2273-4b3f-8234-5205ed2e6e47 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Terminating instance [ 863.502177] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4d4725fd-d007-4a2e-99b6-fa7d7448c81f tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "1bba4036-5211-4a43-b2e6-5276851335e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.946s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.542781] env[61987]: DEBUG nova.network.neutron [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 863.655160] env[61987]: DEBUG nova.compute.manager [req-1ccc18df-ca34-4c8a-b76b-a2846d0da77f req-2604b24f-3e9d-4a15-90ce-19d7e61cb6e0 service nova] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Received event network-vif-plugged-a89e61c6-e262-4b30-a88c-a8731128abb1 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 863.655984] env[61987]: DEBUG oslo_concurrency.lockutils [req-1ccc18df-ca34-4c8a-b76b-a2846d0da77f req-2604b24f-3e9d-4a15-90ce-19d7e61cb6e0 service nova] Acquiring lock "769ecb4d-8fdd-4993-8a29-84c30d1089b2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.655984] env[61987]: DEBUG oslo_concurrency.lockutils [req-1ccc18df-ca34-4c8a-b76b-a2846d0da77f req-2604b24f-3e9d-4a15-90ce-19d7e61cb6e0 service nova] Lock "769ecb4d-8fdd-4993-8a29-84c30d1089b2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.655984] env[61987]: DEBUG oslo_concurrency.lockutils [req-1ccc18df-ca34-4c8a-b76b-a2846d0da77f req-2604b24f-3e9d-4a15-90ce-19d7e61cb6e0 service nova] Lock "769ecb4d-8fdd-4993-8a29-84c30d1089b2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.655984] env[61987]: DEBUG nova.compute.manager [req-1ccc18df-ca34-4c8a-b76b-a2846d0da77f req-2604b24f-3e9d-4a15-90ce-19d7e61cb6e0 service nova] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] No waiting events found dispatching network-vif-plugged-a89e61c6-e262-4b30-a88c-a8731128abb1 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 863.660076] env[61987]: WARNING nova.compute.manager [req-1ccc18df-ca34-4c8a-b76b-a2846d0da77f req-2604b24f-3e9d-4a15-90ce-19d7e61cb6e0 service nova] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Received unexpected event network-vif-plugged-a89e61c6-e262-4b30-a88c-a8731128abb1 for instance with vm_state building and task_state spawning. [ 863.692037] env[61987]: DEBUG nova.network.neutron [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Successfully updated port: a89e61c6-e262-4b30-a88c-a8731128abb1 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 863.771244] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb059a07-9624-4a63-8870-6da85f029d38 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.785777] env[61987]: DEBUG oslo_vmware.api [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061809, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.790427] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d8f79a0-dc9e-40bc-a2c5-a9d59908ab14 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.828437] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6784f3c2-92ff-4395-bdd8-e7e16455e6df {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.839184] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08754b95-aa1a-42c3-81c0-78b66e838153 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.851549] env[61987]: DEBUG nova.compute.provider_tree [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.882264] env[61987]: DEBUG oslo_concurrency.lockutils [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.929210] env[61987]: DEBUG nova.compute.manager [req-dfced8de-7234-4e26-afe0-fb16b1108e89 req-304cf87e-ebce-4fa8-946b-0399bfdb5bb4 service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Received event network-vif-unplugged-4e18ece6-5812-4c5c-827d-b3ee7d399008 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 863.929428] env[61987]: DEBUG oslo_concurrency.lockutils [req-dfced8de-7234-4e26-afe0-fb16b1108e89 req-304cf87e-ebce-4fa8-946b-0399bfdb5bb4 service nova] Acquiring lock "51a36102-795e-47b7-b96a-857e54dc703e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.929781] env[61987]: DEBUG oslo_concurrency.lockutils [req-dfced8de-7234-4e26-afe0-fb16b1108e89 req-304cf87e-ebce-4fa8-946b-0399bfdb5bb4 service nova] Lock "51a36102-795e-47b7-b96a-857e54dc703e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.929894] env[61987]: DEBUG oslo_concurrency.lockutils [req-dfced8de-7234-4e26-afe0-fb16b1108e89 req-304cf87e-ebce-4fa8-946b-0399bfdb5bb4 service nova] Lock "51a36102-795e-47b7-b96a-857e54dc703e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.930047] env[61987]: DEBUG nova.compute.manager [req-dfced8de-7234-4e26-afe0-fb16b1108e89 req-304cf87e-ebce-4fa8-946b-0399bfdb5bb4 service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] No waiting events found dispatching network-vif-unplugged-4e18ece6-5812-4c5c-827d-b3ee7d399008 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 863.930194] env[61987]: WARNING nova.compute.manager [req-dfced8de-7234-4e26-afe0-fb16b1108e89 req-304cf87e-ebce-4fa8-946b-0399bfdb5bb4 service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Received unexpected event network-vif-unplugged-4e18ece6-5812-4c5c-827d-b3ee7d399008 for instance with vm_state shelved_offloaded and task_state None. [ 863.930371] env[61987]: DEBUG nova.compute.manager [req-dfced8de-7234-4e26-afe0-fb16b1108e89 req-304cf87e-ebce-4fa8-946b-0399bfdb5bb4 service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Received event network-changed-4e18ece6-5812-4c5c-827d-b3ee7d399008 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 863.930531] env[61987]: DEBUG nova.compute.manager [req-dfced8de-7234-4e26-afe0-fb16b1108e89 req-304cf87e-ebce-4fa8-946b-0399bfdb5bb4 service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Refreshing instance network info cache due to event network-changed-4e18ece6-5812-4c5c-827d-b3ee7d399008. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 863.930740] env[61987]: DEBUG oslo_concurrency.lockutils [req-dfced8de-7234-4e26-afe0-fb16b1108e89 req-304cf87e-ebce-4fa8-946b-0399bfdb5bb4 service nova] Acquiring lock "refresh_cache-51a36102-795e-47b7-b96a-857e54dc703e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.930942] env[61987]: DEBUG oslo_concurrency.lockutils [req-dfced8de-7234-4e26-afe0-fb16b1108e89 req-304cf87e-ebce-4fa8-946b-0399bfdb5bb4 service nova] Acquired lock "refresh_cache-51a36102-795e-47b7-b96a-857e54dc703e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.931052] env[61987]: DEBUG nova.network.neutron [req-dfced8de-7234-4e26-afe0-fb16b1108e89 req-304cf87e-ebce-4fa8-946b-0399bfdb5bb4 service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Refreshing network info cache for port 4e18ece6-5812-4c5c-827d-b3ee7d399008 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 863.933469] env[61987]: DEBUG nova.network.neutron [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Updating instance_info_cache with network_info: [{"id": "e4942ff0-6c8c-4c29-850d-778e6caa8e44", "address": "fa:16:3e:6c:f8:47", "network": {"id": "56353fe7-1959-4e8d-943c-a1ee1ffc860f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-992817851-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e43b1d992c07429685f7c89f1a8cfc9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4942ff0-6c", "ovs_interfaceid": "e4942ff0-6c8c-4c29-850d-778e6caa8e44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.001035] env[61987]: DEBUG nova.compute.manager [None req-26676183-2273-4b3f-8234-5205ed2e6e47 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 864.001217] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-26676183-2273-4b3f-8234-5205ed2e6e47 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 864.001527] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e16033de-5078-4872-8cce-3521d3783308 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.011514] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-036bf418-b2eb-4e02-abc0-0d8c479d2587 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.041809] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-26676183-2273-4b3f-8234-5205ed2e6e47 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 056c181c-37da-49ac-ae1f-f7f2032c579a could not be found. [ 864.042049] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-26676183-2273-4b3f-8234-5205ed2e6e47 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 864.043362] env[61987]: INFO nova.compute.manager [None req-26676183-2273-4b3f-8234-5205ed2e6e47 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 864.043362] env[61987]: DEBUG oslo.service.loopingcall [None req-26676183-2273-4b3f-8234-5205ed2e6e47 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.043362] env[61987]: DEBUG nova.compute.manager [-] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 864.043362] env[61987]: DEBUG nova.network.neutron [-] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 864.195072] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "refresh_cache-769ecb4d-8fdd-4993-8a29-84c30d1089b2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.195072] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired lock "refresh_cache-769ecb4d-8fdd-4993-8a29-84c30d1089b2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.195268] env[61987]: DEBUG nova.network.neutron [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 864.272080] env[61987]: DEBUG oslo_vmware.api [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061809, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.354587] env[61987]: DEBUG nova.scheduler.client.report [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 864.403160] env[61987]: DEBUG oslo_concurrency.lockutils [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "1bba4036-5211-4a43-b2e6-5276851335e2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.403476] env[61987]: DEBUG oslo_concurrency.lockutils [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "1bba4036-5211-4a43-b2e6-5276851335e2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.403697] env[61987]: DEBUG oslo_concurrency.lockutils [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "1bba4036-5211-4a43-b2e6-5276851335e2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.403890] env[61987]: DEBUG oslo_concurrency.lockutils [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "1bba4036-5211-4a43-b2e6-5276851335e2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.404084] env[61987]: DEBUG oslo_concurrency.lockutils [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "1bba4036-5211-4a43-b2e6-5276851335e2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.408338] env[61987]: INFO nova.compute.manager [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Terminating instance [ 864.437448] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Releasing lock "refresh_cache-455562c7-9fc1-4625-95b0-5eb8c9c8ccaa" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.437448] env[61987]: DEBUG nova.compute.manager [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Instance network_info: |[{"id": "e4942ff0-6c8c-4c29-850d-778e6caa8e44", "address": "fa:16:3e:6c:f8:47", "network": {"id": "56353fe7-1959-4e8d-943c-a1ee1ffc860f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-992817851-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e43b1d992c07429685f7c89f1a8cfc9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4942ff0-6c", "ovs_interfaceid": "e4942ff0-6c8c-4c29-850d-778e6caa8e44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 864.439827] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:f8:47', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '93c5b7ce-4c84-40bc-884c-b2453e0eee69', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e4942ff0-6c8c-4c29-850d-778e6caa8e44', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 864.448122] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Creating folder: Project (e43b1d992c07429685f7c89f1a8cfc9a). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 864.448828] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e774d771-a778-44f3-838d-a66d2c35c2a2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.461959] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Created folder: Project (e43b1d992c07429685f7c89f1a8cfc9a) in parent group-v234219. [ 864.462225] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Creating folder: Instances. Parent ref: group-v234363. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 864.462501] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6ee02991-44f7-44f9-a3a6-24005c707d4f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.473290] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Created folder: Instances in parent group-v234363. [ 864.475478] env[61987]: DEBUG oslo.service.loopingcall [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.475597] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 864.476828] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-36f3fd60-d159-4f29-a0b4-5e3facf4d0db {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.499222] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 864.499222] env[61987]: value = "task-1061829" [ 864.499222] env[61987]: _type = "Task" [ 864.499222] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.508346] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061829, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.687935] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Acquiring lock "fe22fc1b-61fb-4088-b5fe-b84eb4ec4099" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.688465] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Lock "fe22fc1b-61fb-4088-b5fe-b84eb4ec4099" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.688560] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Acquiring lock "fe22fc1b-61fb-4088-b5fe-b84eb4ec4099-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.688962] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Lock "fe22fc1b-61fb-4088-b5fe-b84eb4ec4099-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.689070] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Lock "fe22fc1b-61fb-4088-b5fe-b84eb4ec4099-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.691353] env[61987]: INFO nova.compute.manager [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Terminating instance [ 864.735208] env[61987]: DEBUG nova.network.neutron [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 864.756763] env[61987]: DEBUG nova.network.neutron [req-dfced8de-7234-4e26-afe0-fb16b1108e89 req-304cf87e-ebce-4fa8-946b-0399bfdb5bb4 service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Updated VIF entry in instance network info cache for port 4e18ece6-5812-4c5c-827d-b3ee7d399008. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 864.757177] env[61987]: DEBUG nova.network.neutron [req-dfced8de-7234-4e26-afe0-fb16b1108e89 req-304cf87e-ebce-4fa8-946b-0399bfdb5bb4 service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Updating instance_info_cache with network_info: [{"id": "4e18ece6-5812-4c5c-827d-b3ee7d399008", "address": "fa:16:3e:1a:e6:c9", "network": {"id": "abcd2857-4c5a-4546-a003-858ddb17b0e6", "bridge": null, "label": "tempest-ServersNegativeTestJSON-124838733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3fd98395e57149c987f82585296751b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap4e18ece6-58", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.770301] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "69a22e40-d469-4500-926e-0a12a233f252" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.770593] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "69a22e40-d469-4500-926e-0a12a233f252" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.776485] env[61987]: DEBUG oslo_vmware.api [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061809, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.830208] env[61987]: DEBUG nova.network.neutron [-] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.861101] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.885s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.864798] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.827s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.865203] env[61987]: DEBUG nova.objects.instance [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Lazy-loading 'resources' on Instance uuid 98d895ee-1d1f-49cd-a055-fe6c8462a692 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 864.888935] env[61987]: INFO nova.scheduler.client.report [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Deleted allocations for instance 8f9aab3e-96ee-44c6-8d25-706a0ff857c4 [ 864.911716] env[61987]: DEBUG nova.compute.manager [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 864.911716] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 864.912693] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2a4af63-2800-427c-a0bd-9ec5e61f2970 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.920478] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 864.921158] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b7cac5be-8652-45eb-b2a3-1d557608ffd6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.927540] env[61987]: DEBUG oslo_vmware.api [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 864.927540] env[61987]: value = "task-1061830" [ 864.927540] env[61987]: _type = "Task" [ 864.927540] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.936310] env[61987]: DEBUG oslo_vmware.api [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061830, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.953531] env[61987]: DEBUG nova.network.neutron [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Updating instance_info_cache with network_info: [{"id": "a89e61c6-e262-4b30-a88c-a8731128abb1", "address": "fa:16:3e:83:1f:bc", "network": {"id": "d5ed1e9e-9033-441e-bfcb-2ac643d7f03f", "bridge": "br-int", "label": "tempest-ServersTestJSON-8466993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e62741abfb4471bbc7e11ae7d407ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39ab9baf-90cd-4fe2-8d56-434f8210fc19", "external-id": "nsx-vlan-transportzone-713", "segmentation_id": 713, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa89e61c6-e2", "ovs_interfaceid": "a89e61c6-e262-4b30-a88c-a8731128abb1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.009567] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061829, 'name': CreateVM_Task, 'duration_secs': 0.326138} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.009646] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 865.010370] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.010553] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.010878] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 865.011250] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f09e3e3a-e2dc-4928-918a-25256c573010 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.015794] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for the task: (returnval){ [ 865.015794] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52534cd9-3ee7-3243-81f6-23eab6ffd5f8" [ 865.015794] env[61987]: _type = "Task" [ 865.015794] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.023639] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52534cd9-3ee7-3243-81f6-23eab6ffd5f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.201061] env[61987]: DEBUG nova.compute.manager [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 865.201354] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 865.202337] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d11c54e-f64b-4a80-915a-b20021e31347 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.210364] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 865.210597] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-098de463-67f0-4c96-b5de-1a377debf310 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.216357] env[61987]: DEBUG oslo_vmware.api [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Waiting for the task: (returnval){ [ 865.216357] env[61987]: value = "task-1061831" [ 865.216357] env[61987]: _type = "Task" [ 865.216357] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.223547] env[61987]: DEBUG oslo_vmware.api [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': task-1061831, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.259737] env[61987]: DEBUG oslo_concurrency.lockutils [req-dfced8de-7234-4e26-afe0-fb16b1108e89 req-304cf87e-ebce-4fa8-946b-0399bfdb5bb4 service nova] Releasing lock "refresh_cache-51a36102-795e-47b7-b96a-857e54dc703e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.268844] env[61987]: DEBUG oslo_vmware.api [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061809, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.277376] env[61987]: DEBUG nova.compute.manager [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 865.332888] env[61987]: INFO nova.compute.manager [-] [instance: 056c181c-37da-49ac-ae1f-f7f2032c579a] Took 1.29 seconds to deallocate network for instance. [ 865.395216] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7b48bd21-1d7a-4d85-9b21-43346361dbc6 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "8f9aab3e-96ee-44c6-8d25-706a0ff857c4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.679s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.441826] env[61987]: DEBUG oslo_vmware.api [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061830, 'name': PowerOffVM_Task, 'duration_secs': 0.187747} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.446786] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 865.446786] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 865.449216] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1469e8f2-1730-4ccd-a51c-ba48d7c30468 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.456116] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Releasing lock "refresh_cache-769ecb4d-8fdd-4993-8a29-84c30d1089b2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.456116] env[61987]: DEBUG nova.compute.manager [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Instance network_info: |[{"id": "a89e61c6-e262-4b30-a88c-a8731128abb1", "address": "fa:16:3e:83:1f:bc", "network": {"id": "d5ed1e9e-9033-441e-bfcb-2ac643d7f03f", "bridge": "br-int", "label": "tempest-ServersTestJSON-8466993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e62741abfb4471bbc7e11ae7d407ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39ab9baf-90cd-4fe2-8d56-434f8210fc19", "external-id": "nsx-vlan-transportzone-713", "segmentation_id": 713, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa89e61c6-e2", "ovs_interfaceid": "a89e61c6-e262-4b30-a88c-a8731128abb1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 865.456958] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:1f:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '39ab9baf-90cd-4fe2-8d56-434f8210fc19', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a89e61c6-e262-4b30-a88c-a8731128abb1', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 865.463906] env[61987]: DEBUG oslo.service.loopingcall [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 865.467343] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 865.468686] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2bfd0b21-a62c-4c7d-a00a-bfa8c9080e5d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.490504] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 865.490504] env[61987]: value = "task-1061833" [ 865.490504] env[61987]: _type = "Task" [ 865.490504] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.501652] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061833, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.512227] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 865.512483] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 865.512713] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Deleting the datastore file [datastore2] 1bba4036-5211-4a43-b2e6-5276851335e2 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 865.512956] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f72bb95-1e72-4f61-8cb8-5b8415a94418 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.520648] env[61987]: DEBUG oslo_vmware.api [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for the task: (returnval){ [ 865.520648] env[61987]: value = "task-1061834" [ 865.520648] env[61987]: _type = "Task" [ 865.520648] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.532342] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52534cd9-3ee7-3243-81f6-23eab6ffd5f8, 'name': SearchDatastore_Task, 'duration_secs': 0.009919} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.533247] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.533488] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 865.533751] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.533908] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.534119] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 865.534458] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-22d69632-8e72-413d-9d23-82d57861d992 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.539900] env[61987]: DEBUG oslo_vmware.api [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061834, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.547355] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 865.547355] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 865.547617] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9ca8613-951a-4c86-b504-5da973a80951 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.553475] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for the task: (returnval){ [ 865.553475] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52e1ff2e-78db-4de8-cad1-ba9b0945f87d" [ 865.553475] env[61987]: _type = "Task" [ 865.553475] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.563533] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52e1ff2e-78db-4de8-cad1-ba9b0945f87d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.628630] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-733d74a0-1d80-4aee-98e1-e920e59ffe73 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.637095] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa3f9a0f-31df-4d98-8579-ba8e0737c31c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.670201] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bde3c42-9461-4a7f-aa37-a0a52ac29852 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.678773] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bc461c6-0519-4221-9c53-1eb7d6de0a67 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.693110] env[61987]: DEBUG nova.compute.provider_tree [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.725594] env[61987]: DEBUG oslo_vmware.api [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': task-1061831, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.769784] env[61987]: DEBUG oslo_vmware.api [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061809, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.798961] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.866021] env[61987]: DEBUG nova.compute.manager [req-e8140f78-d732-449d-8386-47b81a2952f1 req-9cc58845-a100-44ec-80a6-02b3faa8994d service nova] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Received event network-changed-a89e61c6-e262-4b30-a88c-a8731128abb1 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 865.866271] env[61987]: DEBUG nova.compute.manager [req-e8140f78-d732-449d-8386-47b81a2952f1 req-9cc58845-a100-44ec-80a6-02b3faa8994d service nova] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Refreshing instance network info cache due to event network-changed-a89e61c6-e262-4b30-a88c-a8731128abb1. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 865.866481] env[61987]: DEBUG oslo_concurrency.lockutils [req-e8140f78-d732-449d-8386-47b81a2952f1 req-9cc58845-a100-44ec-80a6-02b3faa8994d service nova] Acquiring lock "refresh_cache-769ecb4d-8fdd-4993-8a29-84c30d1089b2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.866648] env[61987]: DEBUG oslo_concurrency.lockutils [req-e8140f78-d732-449d-8386-47b81a2952f1 req-9cc58845-a100-44ec-80a6-02b3faa8994d service nova] Acquired lock "refresh_cache-769ecb4d-8fdd-4993-8a29-84c30d1089b2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.866798] env[61987]: DEBUG nova.network.neutron [req-e8140f78-d732-449d-8386-47b81a2952f1 req-9cc58845-a100-44ec-80a6-02b3faa8994d service nova] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Refreshing network info cache for port a89e61c6-e262-4b30-a88c-a8731128abb1 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 865.953588] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "51a36102-795e-47b7-b96a-857e54dc703e" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.992189] env[61987]: DEBUG oslo_concurrency.lockutils [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "30c4fff0-c77e-4c11-aced-a040003a6b10" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.992496] env[61987]: DEBUG oslo_concurrency.lockutils [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "30c4fff0-c77e-4c11-aced-a040003a6b10" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.992707] env[61987]: DEBUG oslo_concurrency.lockutils [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "30c4fff0-c77e-4c11-aced-a040003a6b10-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.992898] env[61987]: DEBUG oslo_concurrency.lockutils [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "30c4fff0-c77e-4c11-aced-a040003a6b10-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.993084] env[61987]: DEBUG oslo_concurrency.lockutils [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "30c4fff0-c77e-4c11-aced-a040003a6b10-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.998855] env[61987]: INFO nova.compute.manager [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Terminating instance [ 866.005123] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061833, 'name': CreateVM_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.031881] env[61987]: DEBUG oslo_vmware.api [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Task: {'id': task-1061834, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.163572} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.032263] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 866.032370] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 866.032556] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 866.032736] env[61987]: INFO nova.compute.manager [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Took 1.12 seconds to destroy the instance on the hypervisor. [ 866.032983] env[61987]: DEBUG oslo.service.loopingcall [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 866.033201] env[61987]: DEBUG nova.compute.manager [-] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 866.033296] env[61987]: DEBUG nova.network.neutron [-] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 866.064148] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52e1ff2e-78db-4de8-cad1-ba9b0945f87d, 'name': SearchDatastore_Task, 'duration_secs': 0.009691} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.065188] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad388ce4-61cf-46e6-b0a9-90e6d386ddf9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.070635] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for the task: (returnval){ [ 866.070635] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5202097c-7095-97e3-5d41-c50c253bb65e" [ 866.070635] env[61987]: _type = "Task" [ 866.070635] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.078719] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5202097c-7095-97e3-5d41-c50c253bb65e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.196165] env[61987]: DEBUG nova.scheduler.client.report [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 866.227855] env[61987]: DEBUG oslo_vmware.api [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': task-1061831, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.275509] env[61987]: DEBUG oslo_vmware.api [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061809, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.357610] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26676183-2273-4b3f-8234-5205ed2e6e47 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "056c181c-37da-49ac-ae1f-f7f2032c579a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.870s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.507181] env[61987]: DEBUG nova.compute.manager [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 866.509011] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 866.509011] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061833, 'name': CreateVM_Task, 'duration_secs': 0.938522} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.509765] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ae4d9d-727b-4a56-a3b3-8c14d6359843 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.513220] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 866.513220] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.513703] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.513703] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 866.515298] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e2737f9-3409-49e5-a3c3-77f6dbdcc021 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.520548] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 866.523038] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8470b84d-321a-4331-a321-dd1152f8f393 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.531339] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "212e4860-751e-43f7-80c6-25b6784e3541" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.531598] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "212e4860-751e-43f7-80c6-25b6784e3541" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.533206] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 866.533206] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]525f41eb-9b96-0bc6-741e-afc0d448f854" [ 866.533206] env[61987]: _type = "Task" [ 866.533206] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.542868] env[61987]: DEBUG oslo_vmware.api [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 866.542868] env[61987]: value = "task-1061835" [ 866.542868] env[61987]: _type = "Task" [ 866.542868] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.549087] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]525f41eb-9b96-0bc6-741e-afc0d448f854, 'name': SearchDatastore_Task, 'duration_secs': 0.009541} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.553045] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.553045] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 866.553249] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.560173] env[61987]: DEBUG oslo_vmware.api [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061835, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.582460] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5202097c-7095-97e3-5d41-c50c253bb65e, 'name': SearchDatastore_Task, 'duration_secs': 0.009026} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.582762] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.583098] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa/455562c7-9fc1-4625-95b0-5eb8c9c8ccaa.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 866.583474] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.583728] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 866.584030] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a2e484d9-e465-4669-9fa4-363c73744ac8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.587146] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c7019500-46de-46db-bec0-1141e65c9db8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.594605] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for the task: (returnval){ [ 866.594605] env[61987]: value = "task-1061836" [ 866.594605] env[61987]: _type = "Task" [ 866.594605] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.599459] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 866.599683] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 866.600794] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08aadd63-3ae3-4c3d-91a9-874368cff5fd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.607280] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061836, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.610271] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 866.610271] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52fbc013-fda1-ea09-c0cb-c4b48bf7b73a" [ 866.610271] env[61987]: _type = "Task" [ 866.610271] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.617644] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52fbc013-fda1-ea09-c0cb-c4b48bf7b73a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.658204] env[61987]: DEBUG nova.network.neutron [req-e8140f78-d732-449d-8386-47b81a2952f1 req-9cc58845-a100-44ec-80a6-02b3faa8994d service nova] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Updated VIF entry in instance network info cache for port a89e61c6-e262-4b30-a88c-a8731128abb1. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 866.658614] env[61987]: DEBUG nova.network.neutron [req-e8140f78-d732-449d-8386-47b81a2952f1 req-9cc58845-a100-44ec-80a6-02b3faa8994d service nova] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Updating instance_info_cache with network_info: [{"id": "a89e61c6-e262-4b30-a88c-a8731128abb1", "address": "fa:16:3e:83:1f:bc", "network": {"id": "d5ed1e9e-9033-441e-bfcb-2ac643d7f03f", "bridge": "br-int", "label": "tempest-ServersTestJSON-8466993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e62741abfb4471bbc7e11ae7d407ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39ab9baf-90cd-4fe2-8d56-434f8210fc19", "external-id": "nsx-vlan-transportzone-713", "segmentation_id": 713, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa89e61c6-e2", "ovs_interfaceid": "a89e61c6-e262-4b30-a88c-a8731128abb1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.702736] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.839s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.705816] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.244s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.729176] env[61987]: DEBUG oslo_vmware.api [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': task-1061831, 'name': PowerOffVM_Task, 'duration_secs': 1.152904} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.729176] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 866.729176] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 866.729399] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-70ce8951-0884-4277-bf35-81ef86d7df0e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.732344] env[61987]: INFO nova.scheduler.client.report [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Deleted allocations for instance 98d895ee-1d1f-49cd-a055-fe6c8462a692 [ 866.773455] env[61987]: DEBUG oslo_vmware.api [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061809, 'name': ReconfigVM_Task, 'duration_secs': 9.044404} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.773455] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.773455] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Reconfigured VM to detach interface {{(pid=61987) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 866.791552] env[61987]: DEBUG nova.network.neutron [-] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.034947] env[61987]: DEBUG nova.compute.manager [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 867.063509] env[61987]: DEBUG oslo_vmware.api [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061835, 'name': PowerOffVM_Task, 'duration_secs': 0.16132} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.063788] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 867.063896] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 867.064965] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5fc20ba4-af3e-4631-ab69-2d9cc4de3e39 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.107425] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061836, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.491372} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.107831] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa/455562c7-9fc1-4625-95b0-5eb8c9c8ccaa.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 867.108190] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 867.108574] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bb58eb75-1886-46ed-9de0-35258383d132 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.119780] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52fbc013-fda1-ea09-c0cb-c4b48bf7b73a, 'name': SearchDatastore_Task, 'duration_secs': 0.00751} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.126583] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for the task: (returnval){ [ 867.126583] env[61987]: value = "task-1061839" [ 867.126583] env[61987]: _type = "Task" [ 867.126583] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.126839] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7034b25-df1d-4061-8472-2cf2a18bb1ac {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.139739] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 867.139739] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52aefd1f-888a-021b-01e1-29648f639809" [ 867.139739] env[61987]: _type = "Task" [ 867.139739] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.142873] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061839, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.152667] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52aefd1f-888a-021b-01e1-29648f639809, 'name': SearchDatastore_Task, 'duration_secs': 0.008748} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.153531] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.153531] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 769ecb4d-8fdd-4993-8a29-84c30d1089b2/769ecb4d-8fdd-4993-8a29-84c30d1089b2.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 867.153680] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2306dafa-3b82-4218-9751-deb1d97ed2ad {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.161194] env[61987]: DEBUG oslo_concurrency.lockutils [req-e8140f78-d732-449d-8386-47b81a2952f1 req-9cc58845-a100-44ec-80a6-02b3faa8994d service nova] Releasing lock "refresh_cache-769ecb4d-8fdd-4993-8a29-84c30d1089b2" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.161644] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 867.161644] env[61987]: value = "task-1061840" [ 867.161644] env[61987]: _type = "Task" [ 867.161644] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.171483] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061840, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.212021] env[61987]: INFO nova.compute.claims [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 867.239870] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dd8c5a2a-7c01-4bfa-956b-e5699fe8cbf4 tempest-ServerRescueTestJSONUnderV235-1865964970 tempest-ServerRescueTestJSONUnderV235-1865964970-project-member] Lock "98d895ee-1d1f-49cd-a055-fe6c8462a692" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.803s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.296960] env[61987]: INFO nova.compute.manager [-] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Took 1.26 seconds to deallocate network for instance. [ 867.556622] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.643925] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061839, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.516083} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.645047] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 867.645303] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9bf71ff-f067-4d96-85b1-d591ebdbb89b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.680941] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa/455562c7-9fc1-4625-95b0-5eb8c9c8ccaa.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.685900] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4ee6a4c-4a90-447c-9c39-2f2fb2e93440 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.707495] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 867.707495] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 867.707495] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Deleting the datastore file [datastore1] fe22fc1b-61fb-4088-b5fe-b84eb4ec4099 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 867.710265] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-74954d81-16d0-429a-a785-e08ff3640391 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.712456] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061840, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.536694} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.714021] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 769ecb4d-8fdd-4993-8a29-84c30d1089b2/769ecb4d-8fdd-4993-8a29-84c30d1089b2.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 867.715658] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 867.715658] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for the task: (returnval){ [ 867.715658] env[61987]: value = "task-1061841" [ 867.715658] env[61987]: _type = "Task" [ 867.715658] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.716164] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bd04fd6e-d7b5-4788-9bc4-c33e0983373c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.726513] env[61987]: INFO nova.compute.resource_tracker [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Updating resource usage from migration 2bef6405-895b-4208-9a03-bb4e85484aab [ 867.732355] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 867.732355] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 867.732355] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Deleting the datastore file [datastore2] 30c4fff0-c77e-4c11-aced-a040003a6b10 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 867.732587] env[61987]: DEBUG oslo_vmware.api [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Waiting for the task: (returnval){ [ 867.732587] env[61987]: value = "task-1061842" [ 867.732587] env[61987]: _type = "Task" [ 867.732587] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.733197] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c2ac11c7-f716-4d65-8ba4-ac823b75e00d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.746493] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061841, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.746493] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 867.746493] env[61987]: value = "task-1061843" [ 867.746493] env[61987]: _type = "Task" [ 867.746493] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.749961] env[61987]: DEBUG oslo_concurrency.lockutils [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "0422be0a-213b-41e9-b850-c823b52a90ca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.750362] env[61987]: DEBUG oslo_concurrency.lockutils [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "0422be0a-213b-41e9-b850-c823b52a90ca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.760171] env[61987]: DEBUG oslo_vmware.api [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 867.760171] env[61987]: value = "task-1061844" [ 867.760171] env[61987]: _type = "Task" [ 867.760171] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.769090] env[61987]: DEBUG oslo_vmware.api [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': task-1061842, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.780720] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061843, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.787713] env[61987]: DEBUG oslo_vmware.api [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061844, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.805924] env[61987]: DEBUG oslo_concurrency.lockutils [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.903978] env[61987]: DEBUG nova.compute.manager [req-cd01bbbd-ff95-47bc-8290-c1e6f29336c7 req-f8a134b1-ef1e-4a34-ada3-758f6ed6a7bc service nova] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Received event network-vif-deleted-1460c9a3-914a-4c65-a2c3-601f1e07788e {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 867.904391] env[61987]: DEBUG nova.compute.manager [req-cd01bbbd-ff95-47bc-8290-c1e6f29336c7 req-f8a134b1-ef1e-4a34-ada3-758f6ed6a7bc service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Received event network-vif-deleted-ebd9819d-a92d-4034-a00b-af37dbb338ab {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 867.904905] env[61987]: INFO nova.compute.manager [req-cd01bbbd-ff95-47bc-8290-c1e6f29336c7 req-f8a134b1-ef1e-4a34-ada3-758f6ed6a7bc service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Neutron deleted interface ebd9819d-a92d-4034-a00b-af37dbb338ab; detaching it from the instance and deleting it from the info cache [ 867.904905] env[61987]: DEBUG nova.network.neutron [req-cd01bbbd-ff95-47bc-8290-c1e6f29336c7 req-f8a134b1-ef1e-4a34-ada3-758f6ed6a7bc service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Updating instance_info_cache with network_info: [{"id": "e7d811c0-d77e-448d-86fe-acdb61c04284", "address": "fa:16:3e:b0:6d:8c", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7d811c0-d7", "ovs_interfaceid": "e7d811c0-d77e-448d-86fe-acdb61c04284", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.013441] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a05e4e0-d121-4edc-ad4f-57f43b41e97d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.022929] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e7e9ba-f54e-492d-91da-826ca6adedbd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.057406] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b326a17c-1c2f-4125-ab39-5293e467b4cf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.066479] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ebdd9cd-7e7a-4dc4-a61c-aaeda020656a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.085257] env[61987]: DEBUG nova.compute.provider_tree [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.112127] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.112127] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.112127] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.112127] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.112127] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.113221] env[61987]: INFO nova.compute.manager [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Terminating instance [ 868.242963] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061841, 'name': ReconfigVM_Task, 'duration_secs': 0.365194} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.243917] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa/455562c7-9fc1-4625-95b0-5eb8c9c8ccaa.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 868.245063] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7f517604-530c-4560-b958-d88aeefac8e3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.250156] env[61987]: DEBUG oslo_vmware.api [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Task: {'id': task-1061842, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199786} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.254700] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 868.255083] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 868.255424] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 868.255724] env[61987]: INFO nova.compute.manager [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Took 3.05 seconds to destroy the instance on the hypervisor. [ 868.256128] env[61987]: DEBUG oslo.service.loopingcall [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 868.257578] env[61987]: DEBUG nova.compute.manager [-] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 868.257804] env[61987]: DEBUG nova.network.neutron [-] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 868.260065] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for the task: (returnval){ [ 868.260065] env[61987]: value = "task-1061845" [ 868.260065] env[61987]: _type = "Task" [ 868.260065] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.268302] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061843, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072976} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.268302] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 868.268818] env[61987]: DEBUG nova.compute.manager [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 868.272606] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c15059e4-bba3-451d-b45e-667ba10e45c8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.283022] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061845, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.290613] env[61987]: DEBUG oslo_vmware.api [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061844, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196558} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.301257] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 868.302191] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 868.302546] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 868.303024] env[61987]: INFO nova.compute.manager [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Took 1.79 seconds to destroy the instance on the hypervisor. [ 868.303247] env[61987]: DEBUG oslo.service.loopingcall [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 868.315070] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 769ecb4d-8fdd-4993-8a29-84c30d1089b2/769ecb4d-8fdd-4993-8a29-84c30d1089b2.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 868.316053] env[61987]: DEBUG nova.compute.manager [-] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 868.316289] env[61987]: DEBUG nova.network.neutron [-] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 868.317960] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4091d8c7-3d32-43bc-8a67-c647a58de16b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.342022] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 868.342022] env[61987]: value = "task-1061846" [ 868.342022] env[61987]: _type = "Task" [ 868.342022] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.350640] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061846, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.408654] env[61987]: DEBUG oslo_concurrency.lockutils [req-cd01bbbd-ff95-47bc-8290-c1e6f29336c7 req-f8a134b1-ef1e-4a34-ada3-758f6ed6a7bc service nova] Acquiring lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.411899] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "refresh_cache-1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.412169] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "refresh_cache-1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.412248] env[61987]: DEBUG nova.network.neutron [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 868.589335] env[61987]: DEBUG nova.scheduler.client.report [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 868.617701] env[61987]: DEBUG nova.compute.manager [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 868.617933] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 868.619085] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a04de02-baa5-4b0e-856c-f3ce784c2747 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.634879] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 868.634879] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7d8fd722-b624-4167-a346-55146995ae8a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.642453] env[61987]: DEBUG oslo_vmware.api [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 868.642453] env[61987]: value = "task-1061847" [ 868.642453] env[61987]: _type = "Task" [ 868.642453] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.652884] env[61987]: DEBUG oslo_vmware.api [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061847, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.776466] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061845, 'name': Rename_Task, 'duration_secs': 0.162642} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.783334] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 868.783334] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c7bb4262-371c-4b70-96d8-f8163e1e976b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.795083] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for the task: (returnval){ [ 868.795083] env[61987]: value = "task-1061848" [ 868.795083] env[61987]: _type = "Task" [ 868.795083] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.806449] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061848, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.811689] env[61987]: DEBUG oslo_concurrency.lockutils [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.854867] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061846, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.038942] env[61987]: DEBUG nova.compute.manager [req-87c1576c-b380-476d-ab7d-9782bbb58d86 req-25a03ac9-8026-4176-b5a3-dfee1d01744b service nova] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Received event network-vif-deleted-979ec444-341c-4296-b3b7-060800602dd1 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 869.038942] env[61987]: INFO nova.compute.manager [req-87c1576c-b380-476d-ab7d-9782bbb58d86 req-25a03ac9-8026-4176-b5a3-dfee1d01744b service nova] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Neutron deleted interface 979ec444-341c-4296-b3b7-060800602dd1; detaching it from the instance and deleting it from the info cache [ 869.038942] env[61987]: DEBUG nova.network.neutron [req-87c1576c-b380-476d-ab7d-9782bbb58d86 req-25a03ac9-8026-4176-b5a3-dfee1d01744b service nova] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.094168] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.388s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.094461] env[61987]: INFO nova.compute.manager [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Migrating [ 869.103208] env[61987]: DEBUG oslo_concurrency.lockutils [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.221s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.106137] env[61987]: DEBUG nova.objects.instance [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lazy-loading 'resources' on Instance uuid 51a36102-795e-47b7-b96a-857e54dc703e {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 869.156366] env[61987]: DEBUG oslo_vmware.api [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061847, 'name': PowerOffVM_Task, 'duration_secs': 0.341371} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.156894] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 869.156894] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 869.157223] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-be9c81ac-708a-4abe-8296-deef7871adfb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.250420] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 869.250928] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 869.250928] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Deleting the datastore file [datastore1] 1f29d5c1-bd01-47dc-9dcf-4e89791120d0 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 869.251115] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-138f33a6-473e-44db-ae90-4f7ab95f2a59 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.260898] env[61987]: DEBUG oslo_vmware.api [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 869.260898] env[61987]: value = "task-1061850" [ 869.260898] env[61987]: _type = "Task" [ 869.260898] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.270601] env[61987]: DEBUG oslo_vmware.api [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061850, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.304966] env[61987]: DEBUG oslo_vmware.api [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061848, 'name': PowerOnVM_Task, 'duration_secs': 0.499765} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.305281] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 869.308736] env[61987]: INFO nova.compute.manager [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Took 9.23 seconds to spawn the instance on the hypervisor. [ 869.308736] env[61987]: DEBUG nova.compute.manager [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 869.308736] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-177091cb-b0f2-4f34-8b70-d9f5fba3182b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.351923] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061846, 'name': ReconfigVM_Task, 'duration_secs': 0.665439} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.352964] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 769ecb4d-8fdd-4993-8a29-84c30d1089b2/769ecb4d-8fdd-4993-8a29-84c30d1089b2.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 869.353616] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f4a7e1b6-60d8-41b6-8bbc-95c79974e086 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.361606] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 869.361606] env[61987]: value = "task-1061851" [ 869.361606] env[61987]: _type = "Task" [ 869.361606] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.372406] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061851, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.403572] env[61987]: DEBUG nova.network.neutron [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Updating instance_info_cache with network_info: [{"id": "e7d811c0-d77e-448d-86fe-acdb61c04284", "address": "fa:16:3e:b0:6d:8c", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.243", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape7d811c0-d7", "ovs_interfaceid": "e7d811c0-d77e-448d-86fe-acdb61c04284", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.448489] env[61987]: DEBUG nova.network.neutron [-] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.542055] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-74cf2980-8989-434d-ab19-981753e193c6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.553175] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e73b6085-e6aa-44f3-9677-d4d19844430e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.574301] env[61987]: DEBUG nova.network.neutron [-] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.591452] env[61987]: DEBUG nova.compute.manager [req-87c1576c-b380-476d-ab7d-9782bbb58d86 req-25a03ac9-8026-4176-b5a3-dfee1d01744b service nova] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Detach interface failed, port_id=979ec444-341c-4296-b3b7-060800602dd1, reason: Instance 30c4fff0-c77e-4c11-aced-a040003a6b10 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 869.608943] env[61987]: DEBUG nova.objects.instance [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lazy-loading 'numa_topology' on Instance uuid 51a36102-795e-47b7-b96a-857e54dc703e {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 869.615053] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "refresh_cache-ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.615272] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired lock "refresh_cache-ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.615486] env[61987]: DEBUG nova.network.neutron [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 869.635393] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquiring lock "93706a54-1fc4-4aed-8807-ec2c3d3694b4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.636251] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Lock "93706a54-1fc4-4aed-8807-ec2c3d3694b4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.636251] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquiring lock "93706a54-1fc4-4aed-8807-ec2c3d3694b4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.636251] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Lock "93706a54-1fc4-4aed-8807-ec2c3d3694b4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.636495] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Lock "93706a54-1fc4-4aed-8807-ec2c3d3694b4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.638484] env[61987]: INFO nova.compute.manager [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Terminating instance [ 869.772172] env[61987]: DEBUG oslo_vmware.api [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061850, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.360777} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.772479] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 869.772679] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 869.773096] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 869.773383] env[61987]: INFO nova.compute.manager [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Took 1.16 seconds to destroy the instance on the hypervisor. [ 869.773655] env[61987]: DEBUG oslo.service.loopingcall [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.773864] env[61987]: DEBUG nova.compute.manager [-] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 869.773959] env[61987]: DEBUG nova.network.neutron [-] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 869.832570] env[61987]: INFO nova.compute.manager [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Took 14.22 seconds to build instance. [ 869.874089] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061851, 'name': Rename_Task, 'duration_secs': 0.140562} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.874641] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 869.876243] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-145fb7f3-c8c4-44f8-80c2-e685e1c53edc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.885216] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 869.885216] env[61987]: value = "task-1061852" [ 869.885216] env[61987]: _type = "Task" [ 869.885216] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.892994] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061852, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.906886] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "refresh_cache-1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.952918] env[61987]: INFO nova.compute.manager [-] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Took 1.64 seconds to deallocate network for instance. [ 869.994405] env[61987]: DEBUG nova.compute.manager [req-f8056c98-2f06-4832-baa8-ab821a1c493d req-b1eab075-0bb2-44ca-be14-efbd7e059b6f service nova] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Received event network-vif-deleted-b4c377b2-8e62-4cde-ac10-05c4cd2bde7d {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 870.076826] env[61987]: INFO nova.compute.manager [-] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Took 1.82 seconds to deallocate network for instance. [ 870.111792] env[61987]: DEBUG nova.objects.base [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Object Instance<51a36102-795e-47b7-b96a-857e54dc703e> lazy-loaded attributes: resources,numa_topology {{(pid=61987) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 870.142666] env[61987]: DEBUG nova.compute.manager [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 870.143789] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 870.144768] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2f4610-3e74-4346-bab6-805bc2f6a0b0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.158222] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 870.158549] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2066fdfc-9379-4615-b5cc-2ce706994de0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.166844] env[61987]: DEBUG oslo_vmware.api [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 870.166844] env[61987]: value = "task-1061853" [ 870.166844] env[61987]: _type = "Task" [ 870.166844] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.186846] env[61987]: DEBUG oslo_vmware.api [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061853, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.338773] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f1310651-2469-4aac-9335-744135e65988 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "455562c7-9fc1-4625-95b0-5eb8c9c8ccaa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.734s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.397025] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061852, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.413322] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c821eef7-2276-4e42-acce-7d47906c1c65 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "interface-1f29d5c1-bd01-47dc-9dcf-4e89791120d0-ebd9819d-a92d-4034-a00b-af37dbb338ab" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 13.781s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.418942] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8fe3b9d-771c-4ef7-aec7-3f7fd0c84516 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.432704] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-805d6d61-de66-41b8-8c3a-e71978d677a7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.468200] env[61987]: DEBUG oslo_concurrency.lockutils [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.469809] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-522495de-0ec6-408d-a3e3-d7109c9692ca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.482328] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-073f45c1-419a-4d3f-afc7-2cc0d5c7792b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.499272] env[61987]: DEBUG nova.compute.provider_tree [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.549715] env[61987]: DEBUG nova.network.neutron [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Updating instance_info_cache with network_info: [{"id": "507c557c-1a2c-4bcb-a8da-93fc1fa91bd9", "address": "fa:16:3e:8c:d5:4a", "network": {"id": "59d8d7d6-fd71-4dec-802e-e0f9e670ff70", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1845227031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "315a2a3450024ed38e27cd7758f392f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap507c557c-1a", "ovs_interfaceid": "507c557c-1a2c-4bcb-a8da-93fc1fa91bd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.586632] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.681086] env[61987]: DEBUG oslo_vmware.api [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061853, 'name': PowerOffVM_Task, 'duration_secs': 0.297045} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.681460] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 870.681607] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 870.681895] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5c1bf6ba-a049-4e7b-9c75-cb213b1449b7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.772235] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 870.772235] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 870.772235] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Deleting the datastore file [datastore1] 93706a54-1fc4-4aed-8807-ec2c3d3694b4 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 870.772235] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-64f0b8b0-fa4e-455e-b7f5-922a23e7e049 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.779905] env[61987]: DEBUG oslo_vmware.api [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for the task: (returnval){ [ 870.779905] env[61987]: value = "task-1061855" [ 870.779905] env[61987]: _type = "Task" [ 870.779905] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.790230] env[61987]: DEBUG oslo_vmware.api [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061855, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.895502] env[61987]: DEBUG oslo_vmware.api [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061852, 'name': PowerOnVM_Task, 'duration_secs': 0.732745} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.895794] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 870.896009] env[61987]: INFO nova.compute.manager [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Took 8.25 seconds to spawn the instance on the hypervisor. [ 870.896204] env[61987]: DEBUG nova.compute.manager [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 870.896987] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb2fdf53-b5a8-4b87-9a12-19919bf82780 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.003280] env[61987]: DEBUG nova.scheduler.client.report [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 871.054180] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Releasing lock "refresh_cache-ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.121990] env[61987]: DEBUG nova.compute.manager [req-3a1cdd5b-ce62-41d2-815c-b2dd4b73f894 req-837cb356-9d30-4d87-a71b-0cbed7bde070 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Received event network-changed-e4942ff0-6c8c-4c29-850d-778e6caa8e44 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 871.122173] env[61987]: DEBUG nova.compute.manager [req-3a1cdd5b-ce62-41d2-815c-b2dd4b73f894 req-837cb356-9d30-4d87-a71b-0cbed7bde070 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Refreshing instance network info cache due to event network-changed-e4942ff0-6c8c-4c29-850d-778e6caa8e44. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 871.122455] env[61987]: DEBUG oslo_concurrency.lockutils [req-3a1cdd5b-ce62-41d2-815c-b2dd4b73f894 req-837cb356-9d30-4d87-a71b-0cbed7bde070 service nova] Acquiring lock "refresh_cache-455562c7-9fc1-4625-95b0-5eb8c9c8ccaa" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.122662] env[61987]: DEBUG oslo_concurrency.lockutils [req-3a1cdd5b-ce62-41d2-815c-b2dd4b73f894 req-837cb356-9d30-4d87-a71b-0cbed7bde070 service nova] Acquired lock "refresh_cache-455562c7-9fc1-4625-95b0-5eb8c9c8ccaa" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.122847] env[61987]: DEBUG nova.network.neutron [req-3a1cdd5b-ce62-41d2-815c-b2dd4b73f894 req-837cb356-9d30-4d87-a71b-0cbed7bde070 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Refreshing network info cache for port e4942ff0-6c8c-4c29-850d-778e6caa8e44 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 871.124386] env[61987]: DEBUG nova.network.neutron [-] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.292071] env[61987]: DEBUG oslo_vmware.api [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061855, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.418233] env[61987]: INFO nova.compute.manager [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Took 13.22 seconds to build instance. [ 871.509573] env[61987]: DEBUG oslo_concurrency.lockutils [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.406s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.512294] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.714s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.513775] env[61987]: INFO nova.compute.claims [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 871.629209] env[61987]: INFO nova.compute.manager [-] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Took 1.86 seconds to deallocate network for instance. [ 871.791194] env[61987]: DEBUG oslo_vmware.api [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061855, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.921432] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ee1005c1-417b-4fa9-bd6a-09af2fea7e34 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "769ecb4d-8fdd-4993-8a29-84c30d1089b2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.737s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.944307] env[61987]: DEBUG nova.network.neutron [req-3a1cdd5b-ce62-41d2-815c-b2dd4b73f894 req-837cb356-9d30-4d87-a71b-0cbed7bde070 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Updated VIF entry in instance network info cache for port e4942ff0-6c8c-4c29-850d-778e6caa8e44. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 871.944795] env[61987]: DEBUG nova.network.neutron [req-3a1cdd5b-ce62-41d2-815c-b2dd4b73f894 req-837cb356-9d30-4d87-a71b-0cbed7bde070 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Updating instance_info_cache with network_info: [{"id": "e4942ff0-6c8c-4c29-850d-778e6caa8e44", "address": "fa:16:3e:6c:f8:47", "network": {"id": "56353fe7-1959-4e8d-943c-a1ee1ffc860f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-992817851-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e43b1d992c07429685f7c89f1a8cfc9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4942ff0-6c", "ovs_interfaceid": "e4942ff0-6c8c-4c29-850d-778e6caa8e44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.028977] env[61987]: DEBUG oslo_concurrency.lockutils [None req-71516b73-3136-4c98-bda2-6ae9c3caea4f tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "51a36102-795e-47b7-b96a-857e54dc703e" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 29.595s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.029982] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "51a36102-795e-47b7-b96a-857e54dc703e" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 6.076s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.030127] env[61987]: INFO nova.compute.manager [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Unshelving [ 872.092550] env[61987]: DEBUG nova.compute.manager [req-5c069dd1-96d2-4db7-94cc-0bd2dc8b1432 req-a8aa8b41-b24f-489b-a52e-0f14760f392a service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Received event network-vif-deleted-e7d811c0-d77e-448d-86fe-acdb61c04284 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 872.136849] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.291901] env[61987]: DEBUG oslo_vmware.api [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Task: {'id': task-1061855, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.204141} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.292195] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 872.292405] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 872.292606] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 872.292787] env[61987]: INFO nova.compute.manager [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Took 2.15 seconds to destroy the instance on the hypervisor. [ 872.293052] env[61987]: DEBUG oslo.service.loopingcall [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.293257] env[61987]: DEBUG nova.compute.manager [-] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 872.293353] env[61987]: DEBUG nova.network.neutron [-] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 872.448285] env[61987]: DEBUG oslo_concurrency.lockutils [req-3a1cdd5b-ce62-41d2-815c-b2dd4b73f894 req-837cb356-9d30-4d87-a71b-0cbed7bde070 service nova] Releasing lock "refresh_cache-455562c7-9fc1-4625-95b0-5eb8c9c8ccaa" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.570603] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dab08c7-6ee9-49f2-951e-11e0c6af9d36 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.592319] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Updating instance 'ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05' progress to 0 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 872.824342] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-432fb1f0-ed58-423e-a8f3-5d3654ea979a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.831735] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ec1e6c-32fc-4299-b0ab-2445f26d6f60 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.835548] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "769ecb4d-8fdd-4993-8a29-84c30d1089b2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.836861] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "769ecb4d-8fdd-4993-8a29-84c30d1089b2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.836861] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "769ecb4d-8fdd-4993-8a29-84c30d1089b2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.836861] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "769ecb4d-8fdd-4993-8a29-84c30d1089b2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.836861] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "769ecb4d-8fdd-4993-8a29-84c30d1089b2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.838532] env[61987]: INFO nova.compute.manager [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Terminating instance [ 872.869519] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e62a12ac-d917-477c-81e6-3d7e19516661 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.878694] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51e5bd9a-51ce-49ea-9739-e113ec67b38f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.895500] env[61987]: DEBUG nova.compute.provider_tree [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.056338] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.102379] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 873.102862] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-813cab89-5be8-4291-93d0-25670b34b402 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.113526] env[61987]: DEBUG oslo_vmware.api [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 873.113526] env[61987]: value = "task-1061856" [ 873.113526] env[61987]: _type = "Task" [ 873.113526] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.124436] env[61987]: DEBUG oslo_vmware.api [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061856, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.153173] env[61987]: DEBUG nova.compute.manager [req-0d3546aa-8900-4985-bf45-7b919647b519 req-17fc52b6-3614-4327-a98d-6059ea3b0b69 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Received event network-changed-e4942ff0-6c8c-4c29-850d-778e6caa8e44 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 873.153173] env[61987]: DEBUG nova.compute.manager [req-0d3546aa-8900-4985-bf45-7b919647b519 req-17fc52b6-3614-4327-a98d-6059ea3b0b69 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Refreshing instance network info cache due to event network-changed-e4942ff0-6c8c-4c29-850d-778e6caa8e44. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 873.153173] env[61987]: DEBUG oslo_concurrency.lockutils [req-0d3546aa-8900-4985-bf45-7b919647b519 req-17fc52b6-3614-4327-a98d-6059ea3b0b69 service nova] Acquiring lock "refresh_cache-455562c7-9fc1-4625-95b0-5eb8c9c8ccaa" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.153173] env[61987]: DEBUG oslo_concurrency.lockutils [req-0d3546aa-8900-4985-bf45-7b919647b519 req-17fc52b6-3614-4327-a98d-6059ea3b0b69 service nova] Acquired lock "refresh_cache-455562c7-9fc1-4625-95b0-5eb8c9c8ccaa" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.153173] env[61987]: DEBUG nova.network.neutron [req-0d3546aa-8900-4985-bf45-7b919647b519 req-17fc52b6-3614-4327-a98d-6059ea3b0b69 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Refreshing network info cache for port e4942ff0-6c8c-4c29-850d-778e6caa8e44 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 873.370041] env[61987]: DEBUG nova.compute.manager [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 873.370321] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 873.371314] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5322eeab-bc96-4dcc-bc3b-7d0d323603bf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.384097] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 873.384416] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fdc255d8-edcc-41f3-ac3a-1bb83a1ba132 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.391940] env[61987]: DEBUG oslo_vmware.api [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 873.391940] env[61987]: value = "task-1061857" [ 873.391940] env[61987]: _type = "Task" [ 873.391940] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.398573] env[61987]: DEBUG nova.scheduler.client.report [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 873.406095] env[61987]: DEBUG oslo_vmware.api [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061857, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.441786] env[61987]: DEBUG nova.network.neutron [-] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.628023] env[61987]: DEBUG oslo_vmware.api [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061856, 'name': PowerOffVM_Task, 'duration_secs': 0.263575} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.628023] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 873.628023] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Updating instance 'ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05' progress to 17 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 873.902649] env[61987]: DEBUG oslo_vmware.api [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061857, 'name': PowerOffVM_Task, 'duration_secs': 0.283159} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.902946] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 873.903148] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 873.903354] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a782e16b-bff0-41d1-b9b3-a8b3c4153c57 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.908024] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.396s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.908500] env[61987]: DEBUG nova.compute.manager [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 873.911221] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.355s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.914983] env[61987]: INFO nova.compute.claims [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 873.921330] env[61987]: DEBUG nova.network.neutron [req-0d3546aa-8900-4985-bf45-7b919647b519 req-17fc52b6-3614-4327-a98d-6059ea3b0b69 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Updated VIF entry in instance network info cache for port e4942ff0-6c8c-4c29-850d-778e6caa8e44. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 873.922064] env[61987]: DEBUG nova.network.neutron [req-0d3546aa-8900-4985-bf45-7b919647b519 req-17fc52b6-3614-4327-a98d-6059ea3b0b69 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Updating instance_info_cache with network_info: [{"id": "e4942ff0-6c8c-4c29-850d-778e6caa8e44", "address": "fa:16:3e:6c:f8:47", "network": {"id": "56353fe7-1959-4e8d-943c-a1ee1ffc860f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-992817851-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e43b1d992c07429685f7c89f1a8cfc9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4942ff0-6c", "ovs_interfaceid": "e4942ff0-6c8c-4c29-850d-778e6caa8e44", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.944735] env[61987]: INFO nova.compute.manager [-] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Took 1.65 seconds to deallocate network for instance. [ 873.978050] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 873.978335] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 873.978500] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Deleting the datastore file [datastore2] 769ecb4d-8fdd-4993-8a29-84c30d1089b2 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 873.979421] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2a590e4e-e410-4229-a4df-1392ac7800bb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.987385] env[61987]: DEBUG oslo_vmware.api [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 873.987385] env[61987]: value = "task-1061859" [ 873.987385] env[61987]: _type = "Task" [ 873.987385] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.997102] env[61987]: DEBUG oslo_vmware.api [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061859, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.118172] env[61987]: DEBUG nova.compute.manager [req-7a279097-8ee6-4a81-a4d5-5fe2a0dc1eb8 req-12f0a4f9-4ad0-44c4-bbfc-94b083ed9e69 service nova] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Received event network-vif-deleted-6fcb037d-fd48-490d-9898-b952ffaad4ef {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 874.135572] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 874.135855] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.136042] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 874.136250] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.136409] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 874.136565] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 874.136850] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 874.137008] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 874.137195] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 874.137371] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 874.137555] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 874.143093] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0cf31043-46dd-4a3e-987d-af9b0b536bfc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.160529] env[61987]: DEBUG oslo_vmware.api [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 874.160529] env[61987]: value = "task-1061860" [ 874.160529] env[61987]: _type = "Task" [ 874.160529] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.170310] env[61987]: DEBUG oslo_vmware.api [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061860, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.416870] env[61987]: DEBUG nova.compute.utils [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 874.420371] env[61987]: DEBUG nova.compute.manager [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 874.420513] env[61987]: DEBUG nova.network.neutron [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 874.424133] env[61987]: DEBUG oslo_concurrency.lockutils [req-0d3546aa-8900-4985-bf45-7b919647b519 req-17fc52b6-3614-4327-a98d-6059ea3b0b69 service nova] Releasing lock "refresh_cache-455562c7-9fc1-4625-95b0-5eb8c9c8ccaa" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.451055] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.469658] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquiring lock "d8b9e172-c3bf-47d9-969b-7f33475d2d60" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.469940] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "d8b9e172-c3bf-47d9-969b-7f33475d2d60" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.481894] env[61987]: DEBUG nova.policy [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2941054f63934469bf4daa6b5a912e14', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '764082416d314c3f92eb83f576aef222', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 874.500385] env[61987]: DEBUG oslo_vmware.api [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061859, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.36376} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.500661] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 874.500871] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 874.501178] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 874.501372] env[61987]: INFO nova.compute.manager [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Took 1.13 seconds to destroy the instance on the hypervisor. [ 874.501918] env[61987]: DEBUG oslo.service.loopingcall [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 874.502519] env[61987]: DEBUG nova.compute.manager [-] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 874.502594] env[61987]: DEBUG nova.network.neutron [-] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 874.674209] env[61987]: DEBUG oslo_vmware.api [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061860, 'name': ReconfigVM_Task, 'duration_secs': 0.19217} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.674571] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Updating instance 'ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05' progress to 33 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 874.921446] env[61987]: DEBUG nova.compute.manager [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 874.972041] env[61987]: DEBUG nova.compute.manager [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 875.015537] env[61987]: DEBUG nova.network.neutron [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Successfully created port: 785edefc-5e84-4a34-97e1-b7c1a08f218e {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 875.177940] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae570776-ef28-4406-acc8-18c6ed7e020c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.182435] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 875.182911] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.183163] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 875.183419] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.183626] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 875.183828] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 875.184098] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 875.184333] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 875.184588] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 875.184782] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 875.185028] env[61987]: DEBUG nova.virt.hardware [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 875.190290] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Reconfiguring VM instance instance-0000004b to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 875.190609] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5acc3102-1d58-418a-97c4-aa86eb7e68ba {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.212306] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-832de4ac-9201-44e6-8f79-fbad18075a04 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.215678] env[61987]: DEBUG oslo_vmware.api [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 875.215678] env[61987]: value = "task-1061861" [ 875.215678] env[61987]: _type = "Task" [ 875.215678] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.246041] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d11843f4-47de-44fa-8eef-51e73dbbe13b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.252743] env[61987]: DEBUG oslo_vmware.api [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061861, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.260393] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f208e4db-38c6-47e8-91ff-39efb1cc0059 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.274542] env[61987]: DEBUG nova.compute.provider_tree [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.414356] env[61987]: DEBUG nova.network.neutron [-] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.494984] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.726620] env[61987]: DEBUG oslo_vmware.api [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061861, 'name': ReconfigVM_Task, 'duration_secs': 0.163679} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.726752] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Reconfigured VM instance instance-0000004b to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 875.727492] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f2ed14a-d945-4569-a485-c06f4f7e50cf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.749590] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05/ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 875.749866] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c5671af-e7da-44bd-8896-ffa3610f83c4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.768899] env[61987]: DEBUG oslo_vmware.api [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 875.768899] env[61987]: value = "task-1061862" [ 875.768899] env[61987]: _type = "Task" [ 875.768899] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.778095] env[61987]: DEBUG nova.scheduler.client.report [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 875.782206] env[61987]: DEBUG oslo_vmware.api [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061862, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.916557] env[61987]: INFO nova.compute.manager [-] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Took 1.41 seconds to deallocate network for instance. [ 875.936463] env[61987]: DEBUG nova.compute.manager [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 875.964858] env[61987]: DEBUG nova.virt.hardware [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 875.965318] env[61987]: DEBUG nova.virt.hardware [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.965578] env[61987]: DEBUG nova.virt.hardware [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 875.965855] env[61987]: DEBUG nova.virt.hardware [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.966179] env[61987]: DEBUG nova.virt.hardware [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 875.966266] env[61987]: DEBUG nova.virt.hardware [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 875.966413] env[61987]: DEBUG nova.virt.hardware [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 875.966621] env[61987]: DEBUG nova.virt.hardware [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 875.966765] env[61987]: DEBUG nova.virt.hardware [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 875.966935] env[61987]: DEBUG nova.virt.hardware [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 875.967129] env[61987]: DEBUG nova.virt.hardware [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 875.968283] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd2d15c-8779-4f26-9d3a-fc46ce7739d5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.976276] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b35201c1-0a73-4cd9-9d27-0003681ef246 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.143552] env[61987]: DEBUG nova.compute.manager [req-930eff5a-87cb-49ba-9caa-737f62253cef req-131a98c0-adf8-43b1-90f5-d5133e19eb42 service nova] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Received event network-vif-deleted-a89e61c6-e262-4b30-a88c-a8731128abb1 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 876.281649] env[61987]: DEBUG oslo_vmware.api [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061862, 'name': ReconfigVM_Task, 'duration_secs': 0.278107} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.282167] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Reconfigured VM instance instance-0000004b to attach disk [datastore2] ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05/ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.282626] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Updating instance 'ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05' progress to 50 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 876.287521] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.376s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.287996] env[61987]: DEBUG nova.compute.manager [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 876.291484] env[61987]: DEBUG oslo_concurrency.lockutils [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.486s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.291706] env[61987]: DEBUG nova.objects.instance [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lazy-loading 'resources' on Instance uuid 1bba4036-5211-4a43-b2e6-5276851335e2 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 876.424119] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.786027] env[61987]: DEBUG nova.network.neutron [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Successfully updated port: 785edefc-5e84-4a34-97e1-b7c1a08f218e {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 876.796114] env[61987]: DEBUG nova.compute.utils [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 876.799777] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd00e80-75f6-43d9-b4f4-cfee64a3423a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.803702] env[61987]: DEBUG nova.compute.manager [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 876.803806] env[61987]: DEBUG nova.network.neutron [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 876.824699] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfc02866-c01c-40b5-84cb-eee822aefe7b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.852518] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Updating instance 'ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05' progress to 67 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 876.858590] env[61987]: DEBUG nova.policy [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3011f049be54b2cb8922d646b94a310', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd828262de7b14d2ab38193d7d34e8f7a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 877.100014] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a94e53bb-9681-4ec0-8a32-5c58611e6eb5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.109188] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd04b36-cc02-4680-94f8-71e4d7a40341 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.143293] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0c4eb18-1a03-4cb7-a32d-16e2813ef122 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.152174] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eab8cb0-3d8e-4479-8070-76099725c3b2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.168812] env[61987]: DEBUG nova.compute.provider_tree [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 877.197863] env[61987]: DEBUG nova.network.neutron [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Successfully created port: ee6aa50c-8e68-48dc-8986-87c78f51a88a {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 877.288277] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.288522] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.288748] env[61987]: DEBUG nova.network.neutron [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 877.304764] env[61987]: DEBUG nova.compute.manager [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 877.418119] env[61987]: DEBUG nova.network.neutron [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Port 507c557c-1a2c-4bcb-a8da-93fc1fa91bd9 binding to destination host cpu-1 is already ACTIVE {{(pid=61987) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 877.672620] env[61987]: DEBUG nova.scheduler.client.report [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 877.823437] env[61987]: DEBUG nova.network.neutron [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 878.042947] env[61987]: DEBUG nova.network.neutron [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updating instance_info_cache with network_info: [{"id": "785edefc-5e84-4a34-97e1-b7c1a08f218e", "address": "fa:16:3e:bf:8b:d7", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap785edefc-5e", "ovs_interfaceid": "785edefc-5e84-4a34-97e1-b7c1a08f218e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.181369] env[61987]: DEBUG oslo_concurrency.lockutils [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.890s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.183924] env[61987]: DEBUG oslo_concurrency.lockutils [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.373s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.185544] env[61987]: INFO nova.compute.claims [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 878.203228] env[61987]: DEBUG nova.compute.manager [req-05e9cbd0-509d-4bdc-bc5c-a417a25fed50 req-149853b9-e362-4f8c-a979-0f91a2f68383 service nova] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Received event network-vif-plugged-785edefc-5e84-4a34-97e1-b7c1a08f218e {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 878.203228] env[61987]: DEBUG oslo_concurrency.lockutils [req-05e9cbd0-509d-4bdc-bc5c-a417a25fed50 req-149853b9-e362-4f8c-a979-0f91a2f68383 service nova] Acquiring lock "69a22e40-d469-4500-926e-0a12a233f252-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.203331] env[61987]: DEBUG oslo_concurrency.lockutils [req-05e9cbd0-509d-4bdc-bc5c-a417a25fed50 req-149853b9-e362-4f8c-a979-0f91a2f68383 service nova] Lock "69a22e40-d469-4500-926e-0a12a233f252-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.203541] env[61987]: DEBUG oslo_concurrency.lockutils [req-05e9cbd0-509d-4bdc-bc5c-a417a25fed50 req-149853b9-e362-4f8c-a979-0f91a2f68383 service nova] Lock "69a22e40-d469-4500-926e-0a12a233f252-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.203836] env[61987]: DEBUG nova.compute.manager [req-05e9cbd0-509d-4bdc-bc5c-a417a25fed50 req-149853b9-e362-4f8c-a979-0f91a2f68383 service nova] [instance: 69a22e40-d469-4500-926e-0a12a233f252] No waiting events found dispatching network-vif-plugged-785edefc-5e84-4a34-97e1-b7c1a08f218e {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 878.203936] env[61987]: WARNING nova.compute.manager [req-05e9cbd0-509d-4bdc-bc5c-a417a25fed50 req-149853b9-e362-4f8c-a979-0f91a2f68383 service nova] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Received unexpected event network-vif-plugged-785edefc-5e84-4a34-97e1-b7c1a08f218e for instance with vm_state building and task_state spawning. [ 878.204127] env[61987]: DEBUG nova.compute.manager [req-05e9cbd0-509d-4bdc-bc5c-a417a25fed50 req-149853b9-e362-4f8c-a979-0f91a2f68383 service nova] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Received event network-changed-785edefc-5e84-4a34-97e1-b7c1a08f218e {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 878.204359] env[61987]: DEBUG nova.compute.manager [req-05e9cbd0-509d-4bdc-bc5c-a417a25fed50 req-149853b9-e362-4f8c-a979-0f91a2f68383 service nova] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Refreshing instance network info cache due to event network-changed-785edefc-5e84-4a34-97e1-b7c1a08f218e. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 878.204557] env[61987]: DEBUG oslo_concurrency.lockutils [req-05e9cbd0-509d-4bdc-bc5c-a417a25fed50 req-149853b9-e362-4f8c-a979-0f91a2f68383 service nova] Acquiring lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.215818] env[61987]: INFO nova.scheduler.client.report [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Deleted allocations for instance 1bba4036-5211-4a43-b2e6-5276851335e2 [ 878.315792] env[61987]: DEBUG nova.compute.manager [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 878.342277] env[61987]: DEBUG nova.virt.hardware [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 878.342611] env[61987]: DEBUG nova.virt.hardware [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 878.342813] env[61987]: DEBUG nova.virt.hardware [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 878.343123] env[61987]: DEBUG nova.virt.hardware [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 878.343332] env[61987]: DEBUG nova.virt.hardware [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 878.343529] env[61987]: DEBUG nova.virt.hardware [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 878.343777] env[61987]: DEBUG nova.virt.hardware [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 878.343988] env[61987]: DEBUG nova.virt.hardware [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 878.344599] env[61987]: DEBUG nova.virt.hardware [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 878.344599] env[61987]: DEBUG nova.virt.hardware [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 878.344599] env[61987]: DEBUG nova.virt.hardware [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 878.345592] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b257d6-af1c-4cd1-9846-5f1dfc510c76 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.355606] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c211ccbd-c274-4774-95dc-42a0fb6dc037 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.440038] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.440347] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.440541] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.544943] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.545345] env[61987]: DEBUG nova.compute.manager [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Instance network_info: |[{"id": "785edefc-5e84-4a34-97e1-b7c1a08f218e", "address": "fa:16:3e:bf:8b:d7", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap785edefc-5e", "ovs_interfaceid": "785edefc-5e84-4a34-97e1-b7c1a08f218e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 878.545698] env[61987]: DEBUG oslo_concurrency.lockutils [req-05e9cbd0-509d-4bdc-bc5c-a417a25fed50 req-149853b9-e362-4f8c-a979-0f91a2f68383 service nova] Acquired lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.545895] env[61987]: DEBUG nova.network.neutron [req-05e9cbd0-509d-4bdc-bc5c-a417a25fed50 req-149853b9-e362-4f8c-a979-0f91a2f68383 service nova] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Refreshing network info cache for port 785edefc-5e84-4a34-97e1-b7c1a08f218e {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 878.547238] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bf:8b:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2a111da1-447a-4722-9662-b0d6a2886a65', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '785edefc-5e84-4a34-97e1-b7c1a08f218e', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 878.555171] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Creating folder: Project (764082416d314c3f92eb83f576aef222). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 878.556333] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-93097c46-7e5d-4c03-b742-3a3849e8757f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.571025] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Created folder: Project (764082416d314c3f92eb83f576aef222) in parent group-v234219. [ 878.571262] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Creating folder: Instances. Parent ref: group-v234367. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 878.571538] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0abffc96-45e3-4e01-831b-037b15b46f24 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.584023] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Created folder: Instances in parent group-v234367. [ 878.584398] env[61987]: DEBUG oslo.service.loopingcall [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.584633] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 878.584820] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c5a0567b-557b-40f5-9f63-d76821be7e6c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.606056] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 878.606056] env[61987]: value = "task-1061865" [ 878.606056] env[61987]: _type = "Task" [ 878.606056] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.615028] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061865, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.723258] env[61987]: DEBUG oslo_concurrency.lockutils [None req-15298582-6a34-44e0-bf5d-fb2ef2c99d77 tempest-ImagesOneServerNegativeTestJSON-874382516 tempest-ImagesOneServerNegativeTestJSON-874382516-project-member] Lock "1bba4036-5211-4a43-b2e6-5276851335e2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.320s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.118088] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061865, 'name': CreateVM_Task, 'duration_secs': 0.351128} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.118447] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 879.119784] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.120111] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.120687] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 879.122019] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-520f56e9-dae1-416c-b2a1-154757edf666 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.130612] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 879.130612] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52118de5-8e92-2373-a2ae-3d003fa3e4cf" [ 879.130612] env[61987]: _type = "Task" [ 879.130612] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.143386] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52118de5-8e92-2373-a2ae-3d003fa3e4cf, 'name': SearchDatastore_Task, 'duration_secs': 0.012239} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.143953] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.144388] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 879.144801] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.145091] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.145409] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 879.145793] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3049c88b-c604-4012-ab6f-079fae9d8f22 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.158428] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 879.158761] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 879.159607] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5da5e6e-d487-463e-b5d8-9985967aafb2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.166111] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 879.166111] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5248e282-31b0-0e62-6ac0-4607437d4afb" [ 879.166111] env[61987]: _type = "Task" [ 879.166111] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.176223] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5248e282-31b0-0e62-6ac0-4607437d4afb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.349105] env[61987]: DEBUG nova.network.neutron [req-05e9cbd0-509d-4bdc-bc5c-a417a25fed50 req-149853b9-e362-4f8c-a979-0f91a2f68383 service nova] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updated VIF entry in instance network info cache for port 785edefc-5e84-4a34-97e1-b7c1a08f218e. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 879.349492] env[61987]: DEBUG nova.network.neutron [req-05e9cbd0-509d-4bdc-bc5c-a417a25fed50 req-149853b9-e362-4f8c-a979-0f91a2f68383 service nova] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updating instance_info_cache with network_info: [{"id": "785edefc-5e84-4a34-97e1-b7c1a08f218e", "address": "fa:16:3e:bf:8b:d7", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap785edefc-5e", "ovs_interfaceid": "785edefc-5e84-4a34-97e1-b7c1a08f218e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.470152] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40f2a117-752d-4070-bcf7-2e72a35f339b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.479279] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bb4add7-a15f-4f87-a6a1-b7082c6512bb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.511765] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "refresh_cache-ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.511982] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired lock "refresh_cache-ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.512154] env[61987]: DEBUG nova.network.neutron [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 879.514581] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-179f7521-24c5-4f0d-a960-9a08a77c56fa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.523873] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-438df67e-3e3f-4799-ada9-e5088801ef59 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.545624] env[61987]: DEBUG nova.compute.provider_tree [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.678988] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5248e282-31b0-0e62-6ac0-4607437d4afb, 'name': SearchDatastore_Task, 'duration_secs': 0.011312} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.679842] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1bd96a34-b0a5-4c42-937b-72614cfeaebd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.686521] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 879.686521] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]529c03a5-96f3-4896-28d2-00ff13152c44" [ 879.686521] env[61987]: _type = "Task" [ 879.686521] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.695509] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]529c03a5-96f3-4896-28d2-00ff13152c44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.768032] env[61987]: DEBUG nova.network.neutron [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Successfully updated port: ee6aa50c-8e68-48dc-8986-87c78f51a88a {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 879.853417] env[61987]: DEBUG oslo_concurrency.lockutils [req-05e9cbd0-509d-4bdc-bc5c-a417a25fed50 req-149853b9-e362-4f8c-a979-0f91a2f68383 service nova] Releasing lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.050037] env[61987]: DEBUG nova.scheduler.client.report [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 880.198399] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]529c03a5-96f3-4896-28d2-00ff13152c44, 'name': SearchDatastore_Task, 'duration_secs': 0.016681} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.198832] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.199372] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 69a22e40-d469-4500-926e-0a12a233f252/69a22e40-d469-4500-926e-0a12a233f252.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 880.202428] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7d411a8b-9793-4e71-9ee9-e8dc04ecd476 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.211155] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 880.211155] env[61987]: value = "task-1061866" [ 880.211155] env[61987]: _type = "Task" [ 880.211155] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.219857] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1061866, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.229663] env[61987]: DEBUG nova.compute.manager [req-54279a8a-0c86-4f4d-8047-8c68cf27e92e req-adc433a8-a4ad-41f2-b114-c1722f100412 service nova] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Received event network-vif-plugged-ee6aa50c-8e68-48dc-8986-87c78f51a88a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 880.229898] env[61987]: DEBUG oslo_concurrency.lockutils [req-54279a8a-0c86-4f4d-8047-8c68cf27e92e req-adc433a8-a4ad-41f2-b114-c1722f100412 service nova] Acquiring lock "212e4860-751e-43f7-80c6-25b6784e3541-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.230162] env[61987]: DEBUG oslo_concurrency.lockutils [req-54279a8a-0c86-4f4d-8047-8c68cf27e92e req-adc433a8-a4ad-41f2-b114-c1722f100412 service nova] Lock "212e4860-751e-43f7-80c6-25b6784e3541-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.230349] env[61987]: DEBUG oslo_concurrency.lockutils [req-54279a8a-0c86-4f4d-8047-8c68cf27e92e req-adc433a8-a4ad-41f2-b114-c1722f100412 service nova] Lock "212e4860-751e-43f7-80c6-25b6784e3541-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.230561] env[61987]: DEBUG nova.compute.manager [req-54279a8a-0c86-4f4d-8047-8c68cf27e92e req-adc433a8-a4ad-41f2-b114-c1722f100412 service nova] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] No waiting events found dispatching network-vif-plugged-ee6aa50c-8e68-48dc-8986-87c78f51a88a {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 880.230783] env[61987]: WARNING nova.compute.manager [req-54279a8a-0c86-4f4d-8047-8c68cf27e92e req-adc433a8-a4ad-41f2-b114-c1722f100412 service nova] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Received unexpected event network-vif-plugged-ee6aa50c-8e68-48dc-8986-87c78f51a88a for instance with vm_state building and task_state spawning. [ 880.230994] env[61987]: DEBUG nova.compute.manager [req-54279a8a-0c86-4f4d-8047-8c68cf27e92e req-adc433a8-a4ad-41f2-b114-c1722f100412 service nova] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Received event network-changed-ee6aa50c-8e68-48dc-8986-87c78f51a88a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 880.231251] env[61987]: DEBUG nova.compute.manager [req-54279a8a-0c86-4f4d-8047-8c68cf27e92e req-adc433a8-a4ad-41f2-b114-c1722f100412 service nova] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Refreshing instance network info cache due to event network-changed-ee6aa50c-8e68-48dc-8986-87c78f51a88a. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 880.231453] env[61987]: DEBUG oslo_concurrency.lockutils [req-54279a8a-0c86-4f4d-8047-8c68cf27e92e req-adc433a8-a4ad-41f2-b114-c1722f100412 service nova] Acquiring lock "refresh_cache-212e4860-751e-43f7-80c6-25b6784e3541" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.231601] env[61987]: DEBUG oslo_concurrency.lockutils [req-54279a8a-0c86-4f4d-8047-8c68cf27e92e req-adc433a8-a4ad-41f2-b114-c1722f100412 service nova] Acquired lock "refresh_cache-212e4860-751e-43f7-80c6-25b6784e3541" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.231763] env[61987]: DEBUG nova.network.neutron [req-54279a8a-0c86-4f4d-8047-8c68cf27e92e req-adc433a8-a4ad-41f2-b114-c1722f100412 service nova] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Refreshing network info cache for port ee6aa50c-8e68-48dc-8986-87c78f51a88a {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 880.271253] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "refresh_cache-212e4860-751e-43f7-80c6-25b6784e3541" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.281823] env[61987]: DEBUG nova.network.neutron [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Updating instance_info_cache with network_info: [{"id": "507c557c-1a2c-4bcb-a8da-93fc1fa91bd9", "address": "fa:16:3e:8c:d5:4a", "network": {"id": "59d8d7d6-fd71-4dec-802e-e0f9e670ff70", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1845227031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "315a2a3450024ed38e27cd7758f392f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap507c557c-1a", "ovs_interfaceid": "507c557c-1a2c-4bcb-a8da-93fc1fa91bd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.558271] env[61987]: DEBUG oslo_concurrency.lockutils [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.374s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.558787] env[61987]: DEBUG nova.compute.manager [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 880.561670] env[61987]: DEBUG oslo_concurrency.lockutils [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.094s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.561938] env[61987]: DEBUG nova.objects.instance [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lazy-loading 'resources' on Instance uuid 30c4fff0-c77e-4c11-aced-a040003a6b10 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 880.722664] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1061866, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.781677] env[61987]: DEBUG nova.network.neutron [req-54279a8a-0c86-4f4d-8047-8c68cf27e92e req-adc433a8-a4ad-41f2-b114-c1722f100412 service nova] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 880.785189] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Releasing lock "refresh_cache-ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.894602] env[61987]: DEBUG nova.network.neutron [req-54279a8a-0c86-4f4d-8047-8c68cf27e92e req-adc433a8-a4ad-41f2-b114-c1722f100412 service nova] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.068176] env[61987]: DEBUG nova.compute.utils [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 881.069700] env[61987]: DEBUG nova.compute.manager [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 881.069871] env[61987]: DEBUG nova.network.neutron [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 881.122177] env[61987]: DEBUG nova.policy [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c124c6b758543a68e141796b585e41f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bd47eb44bc334bf3ae5813905903ecbc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 881.225811] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1061866, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.526388} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.228477] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 69a22e40-d469-4500-926e-0a12a233f252/69a22e40-d469-4500-926e-0a12a233f252.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 881.228816] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 881.229812] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6b8468ef-23db-4ff0-84a9-b3eb67a0149f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.237444] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 881.237444] env[61987]: value = "task-1061867" [ 881.237444] env[61987]: _type = "Task" [ 881.237444] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.250854] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1061867, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.310347] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0dfb0d5-70e6-4d0f-852c-a118b4bc83e7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.314184] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40996ddf-35e4-41b0-b076-a6f170c37a2e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.335343] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-103f23b0-9ed8-4f37-b737-5e8e84aa29c8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.339102] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09619188-67b0-46be-8d51-dd07e6d021bc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.347044] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Updating instance 'ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05' progress to 83 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 881.381699] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20a4375-2adc-4f18-bb21-5083d69efcbd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.391931] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c30e8d7f-b4c2-47f6-955a-f93e91207fe9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.396787] env[61987]: DEBUG oslo_concurrency.lockutils [req-54279a8a-0c86-4f4d-8047-8c68cf27e92e req-adc433a8-a4ad-41f2-b114-c1722f100412 service nova] Releasing lock "refresh_cache-212e4860-751e-43f7-80c6-25b6784e3541" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.397183] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquired lock "refresh_cache-212e4860-751e-43f7-80c6-25b6784e3541" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.397339] env[61987]: DEBUG nova.network.neutron [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 881.399745] env[61987]: DEBUG nova.network.neutron [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Successfully created port: 4a1063a6-ed46-438e-af9b-6345483b1c9a {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 881.411207] env[61987]: DEBUG nova.compute.provider_tree [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.572774] env[61987]: DEBUG nova.compute.manager [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 881.748529] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1061867, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069643} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.748857] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 881.749692] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32c2d82a-989a-4ea5-9dd3-542caedd9ef7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.773885] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] 69a22e40-d469-4500-926e-0a12a233f252/69a22e40-d469-4500-926e-0a12a233f252.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 881.774085] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45ca49c4-2258-4202-a90d-52fc34d6c432 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.795985] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 881.795985] env[61987]: value = "task-1061868" [ 881.795985] env[61987]: _type = "Task" [ 881.795985] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.804496] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1061868, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.883177] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 881.883894] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-92e4b45b-d3df-429d-ae8b-e367a1bddf73 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.892170] env[61987]: DEBUG oslo_vmware.api [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 881.892170] env[61987]: value = "task-1061869" [ 881.892170] env[61987]: _type = "Task" [ 881.892170] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.900757] env[61987]: DEBUG oslo_vmware.api [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061869, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.913894] env[61987]: DEBUG nova.scheduler.client.report [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 881.937986] env[61987]: DEBUG nova.network.neutron [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 882.072507] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 882.072825] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 882.117991] env[61987]: DEBUG nova.network.neutron [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Updating instance_info_cache with network_info: [{"id": "ee6aa50c-8e68-48dc-8986-87c78f51a88a", "address": "fa:16:3e:44:05:2f", "network": {"id": "e2d0ce90-ff14-47de-bc4e-a2564d1268a7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-222691221-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d828262de7b14d2ab38193d7d34e8f7a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee6aa50c-8e", "ovs_interfaceid": "ee6aa50c-8e68-48dc-8986-87c78f51a88a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.306695] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1061868, 'name': ReconfigVM_Task, 'duration_secs': 0.342157} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.307031] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Reconfigured VM instance instance-0000004f to attach disk [datastore2] 69a22e40-d469-4500-926e-0a12a233f252/69a22e40-d469-4500-926e-0a12a233f252.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 882.307676] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-124f2970-52a4-45c9-b097-1fb98a3b39bb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.314845] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 882.314845] env[61987]: value = "task-1061870" [ 882.314845] env[61987]: _type = "Task" [ 882.314845] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.323538] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1061870, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.403018] env[61987]: DEBUG oslo_vmware.api [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061869, 'name': PowerOnVM_Task, 'duration_secs': 0.412379} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.403323] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 882.403525] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2adf4654-81ae-4fbf-ae67-2d100d886691 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Updating instance 'ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05' progress to 100 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 882.422332] env[61987]: DEBUG oslo_concurrency.lockutils [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.861s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.424771] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.838s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.425017] env[61987]: DEBUG nova.objects.instance [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Lazy-loading 'resources' on Instance uuid fe22fc1b-61fb-4088-b5fe-b84eb4ec4099 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 882.452924] env[61987]: INFO nova.scheduler.client.report [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Deleted allocations for instance 30c4fff0-c77e-4c11-aced-a040003a6b10 [ 882.581162] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 882.581533] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Starting heal instance info cache {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10401}} [ 882.583538] env[61987]: DEBUG nova.compute.manager [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 882.618949] env[61987]: DEBUG nova.virt.hardware [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 882.619328] env[61987]: DEBUG nova.virt.hardware [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 882.619556] env[61987]: DEBUG nova.virt.hardware [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 882.619791] env[61987]: DEBUG nova.virt.hardware [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 882.619964] env[61987]: DEBUG nova.virt.hardware [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 882.620139] env[61987]: DEBUG nova.virt.hardware [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 882.623693] env[61987]: DEBUG nova.virt.hardware [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 882.623693] env[61987]: DEBUG nova.virt.hardware [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 882.623693] env[61987]: DEBUG nova.virt.hardware [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 882.623693] env[61987]: DEBUG nova.virt.hardware [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 882.623693] env[61987]: DEBUG nova.virt.hardware [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 882.623693] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Releasing lock "refresh_cache-212e4860-751e-43f7-80c6-25b6784e3541" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.623693] env[61987]: DEBUG nova.compute.manager [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Instance network_info: |[{"id": "ee6aa50c-8e68-48dc-8986-87c78f51a88a", "address": "fa:16:3e:44:05:2f", "network": {"id": "e2d0ce90-ff14-47de-bc4e-a2564d1268a7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-222691221-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d828262de7b14d2ab38193d7d34e8f7a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f52a458-d157-48a3-b4e2-b8cc0779afe2", "external-id": "nsx-vlan-transportzone-403", "segmentation_id": 403, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee6aa50c-8e", "ovs_interfaceid": "ee6aa50c-8e68-48dc-8986-87c78f51a88a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 882.623693] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c899879-38b6-42cb-99c1-1b2e1c43ded7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.628302] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:44:05:2f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f52a458-d157-48a3-b4e2-b8cc0779afe2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ee6aa50c-8e68-48dc-8986-87c78f51a88a', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 882.634181] env[61987]: DEBUG oslo.service.loopingcall [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.634730] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 882.635526] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-14ed9384-0c29-4c4b-abbc-ef93fa5a7991 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.655484] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a03c3eb-0744-4f3d-95c1-48091e62a71d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.661085] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 882.661085] env[61987]: value = "task-1061871" [ 882.661085] env[61987]: _type = "Task" [ 882.661085] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.681355] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061871, 'name': CreateVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.825605] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1061870, 'name': Rename_Task, 'duration_secs': 0.205456} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.825816] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 882.826305] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ecdb95cd-512e-44e5-bbc5-c342a27e9b85 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.836424] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 882.836424] env[61987]: value = "task-1061872" [ 882.836424] env[61987]: _type = "Task" [ 882.836424] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.847471] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1061872, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.962766] env[61987]: DEBUG oslo_concurrency.lockutils [None req-089e4b7e-ee7c-4324-8131-800284fe6f75 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "30c4fff0-c77e-4c11-aced-a040003a6b10" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.970s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.022042] env[61987]: DEBUG nova.compute.manager [req-10fe68ac-a8e3-4321-a9d5-4c7e77343deb req-f1da639b-eb6d-4ce2-893b-2e0f6cbfe4c4 service nova] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Received event network-vif-plugged-4a1063a6-ed46-438e-af9b-6345483b1c9a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 883.022191] env[61987]: DEBUG oslo_concurrency.lockutils [req-10fe68ac-a8e3-4321-a9d5-4c7e77343deb req-f1da639b-eb6d-4ce2-893b-2e0f6cbfe4c4 service nova] Acquiring lock "0422be0a-213b-41e9-b850-c823b52a90ca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.022437] env[61987]: DEBUG oslo_concurrency.lockutils [req-10fe68ac-a8e3-4321-a9d5-4c7e77343deb req-f1da639b-eb6d-4ce2-893b-2e0f6cbfe4c4 service nova] Lock "0422be0a-213b-41e9-b850-c823b52a90ca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.022623] env[61987]: DEBUG oslo_concurrency.lockutils [req-10fe68ac-a8e3-4321-a9d5-4c7e77343deb req-f1da639b-eb6d-4ce2-893b-2e0f6cbfe4c4 service nova] Lock "0422be0a-213b-41e9-b850-c823b52a90ca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.022813] env[61987]: DEBUG nova.compute.manager [req-10fe68ac-a8e3-4321-a9d5-4c7e77343deb req-f1da639b-eb6d-4ce2-893b-2e0f6cbfe4c4 service nova] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] No waiting events found dispatching network-vif-plugged-4a1063a6-ed46-438e-af9b-6345483b1c9a {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 883.023661] env[61987]: WARNING nova.compute.manager [req-10fe68ac-a8e3-4321-a9d5-4c7e77343deb req-f1da639b-eb6d-4ce2-893b-2e0f6cbfe4c4 service nova] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Received unexpected event network-vif-plugged-4a1063a6-ed46-438e-af9b-6345483b1c9a for instance with vm_state building and task_state spawning. [ 883.118045] env[61987]: DEBUG nova.network.neutron [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Successfully updated port: 4a1063a6-ed46-438e-af9b-6345483b1c9a {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 883.171254] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061871, 'name': CreateVM_Task, 'duration_secs': 0.491122} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.171521] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 883.172649] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.172908] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.173360] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 883.176360] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eda4fb90-09bf-4767-a907-6e16ec5cdcc6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.182080] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 883.182080] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52ef9c33-212f-1586-16ca-a261df15c659" [ 883.182080] env[61987]: _type = "Task" [ 883.182080] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.195826] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52ef9c33-212f-1586-16ca-a261df15c659, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.214206] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6ddef22-a96b-4d48-9126-6bd98720c66c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.222797] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d115e695-345c-4fb3-9204-3c36e3bbdd59 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.257459] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-113cf4d4-9f47-4784-b0b2-1d60b8c22042 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.267790] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b69948-ade1-4db7-a110-d8647063c347 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.286361] env[61987]: DEBUG nova.compute.provider_tree [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 883.346647] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1061872, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.631936] env[61987]: DEBUG oslo_concurrency.lockutils [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "refresh_cache-0422be0a-213b-41e9-b850-c823b52a90ca" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.632181] env[61987]: DEBUG oslo_concurrency.lockutils [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired lock "refresh_cache-0422be0a-213b-41e9-b850-c823b52a90ca" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.632290] env[61987]: DEBUG nova.network.neutron [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 883.696124] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52ef9c33-212f-1586-16ca-a261df15c659, 'name': SearchDatastore_Task, 'duration_secs': 0.013752} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.696514] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.696699] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 883.696942] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.697140] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.697292] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 883.697560] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-60c0733c-9bfa-4837-a24a-ab7653ac4781 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.707750] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 883.707750] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 883.708183] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-626f778b-c364-43bb-bde0-962c5ad263ed {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.718753] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 883.718753] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]527a11be-5971-9149-15e9-c0d064c51f28" [ 883.718753] env[61987]: _type = "Task" [ 883.718753] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.730057] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]527a11be-5971-9149-15e9-c0d064c51f28, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.791140] env[61987]: DEBUG nova.scheduler.client.report [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 883.849081] env[61987]: DEBUG oslo_vmware.api [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1061872, 'name': PowerOnVM_Task, 'duration_secs': 0.519498} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.849281] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 883.849487] env[61987]: INFO nova.compute.manager [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Took 7.91 seconds to spawn the instance on the hypervisor. [ 883.849670] env[61987]: DEBUG nova.compute.manager [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 883.850551] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d93bc42-55bd-4907-b926-c12645c7d639 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.164661] env[61987]: DEBUG nova.network.neutron [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 884.232277] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]527a11be-5971-9149-15e9-c0d064c51f28, 'name': SearchDatastore_Task, 'duration_secs': 0.009968} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.232277] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd00b7bd-bc9f-4e48-bb1c-d1f8de03ccd9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.238110] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 884.238110] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]520866df-1220-a4f7-8fff-1e84eba39e4a" [ 884.238110] env[61987]: _type = "Task" [ 884.238110] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.247191] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]520866df-1220-a4f7-8fff-1e84eba39e4a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.297144] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.871s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.298382] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.162s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.302280] env[61987]: DEBUG nova.objects.instance [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lazy-loading 'resources' on Instance uuid 1f29d5c1-bd01-47dc-9dcf-4e89791120d0 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 884.333699] env[61987]: INFO nova.scheduler.client.report [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Deleted allocations for instance fe22fc1b-61fb-4088-b5fe-b84eb4ec4099 [ 884.369281] env[61987]: INFO nova.compute.manager [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Took 18.59 seconds to build instance. [ 884.417360] env[61987]: DEBUG oslo_concurrency.lockutils [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "742297eb-c45d-4f45-85d2-e9bdd3106d1e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.417618] env[61987]: DEBUG oslo_concurrency.lockutils [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "742297eb-c45d-4f45-85d2-e9bdd3106d1e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.417882] env[61987]: DEBUG oslo_concurrency.lockutils [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "742297eb-c45d-4f45-85d2-e9bdd3106d1e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.418093] env[61987]: DEBUG oslo_concurrency.lockutils [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "742297eb-c45d-4f45-85d2-e9bdd3106d1e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.418284] env[61987]: DEBUG oslo_concurrency.lockutils [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "742297eb-c45d-4f45-85d2-e9bdd3106d1e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.421027] env[61987]: DEBUG nova.network.neutron [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Updating instance_info_cache with network_info: [{"id": "4a1063a6-ed46-438e-af9b-6345483b1c9a", "address": "fa:16:3e:62:e1:24", "network": {"id": "ca68bf96-3f18-45ee-80fb-64a1260ed996", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-733595377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd47eb44bc334bf3ae5813905903ecbc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a27fd90b-16a5-43af-bede-ae36762ece00", "external-id": "nsx-vlan-transportzone-197", "segmentation_id": 197, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a1063a6-ed", "ovs_interfaceid": "4a1063a6-ed46-438e-af9b-6345483b1c9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.422420] env[61987]: INFO nova.compute.manager [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Terminating instance [ 884.602015] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Skipping network cache update for instance because it has been migrated to another host. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10445}} [ 884.750166] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]520866df-1220-a4f7-8fff-1e84eba39e4a, 'name': SearchDatastore_Task, 'duration_secs': 0.030665} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.750477] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.750727] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 212e4860-751e-43f7-80c6-25b6784e3541/212e4860-751e-43f7-80c6-25b6784e3541.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 884.750997] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a268245e-2d07-46b7-9e7f-140238db81ca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.762679] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 884.762679] env[61987]: value = "task-1061873" [ 884.762679] env[61987]: _type = "Task" [ 884.762679] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.773723] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061873, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.834945] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8d96470-8adc-4093-b2a3-f7b6ae6128cb tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.835950] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8d96470-8adc-4093-b2a3-f7b6ae6128cb tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.836190] env[61987]: DEBUG nova.compute.manager [None req-b8d96470-8adc-4093-b2a3-f7b6ae6128cb tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Going to confirm migration 3 {{(pid=61987) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5249}} [ 884.843234] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e130538f-563f-4f99-85a0-c88d72dfe31e tempest-ServersV294TestFqdnHostnames-217133186 tempest-ServersV294TestFqdnHostnames-217133186-project-member] Lock "fe22fc1b-61fb-4088-b5fe-b84eb4ec4099" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.155s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.873455] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ea48c55-30ef-45b6-a12d-bd36206acd72 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "69a22e40-d469-4500-926e-0a12a233f252" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.103s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.926324] env[61987]: DEBUG oslo_concurrency.lockutils [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Releasing lock "refresh_cache-0422be0a-213b-41e9-b850-c823b52a90ca" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.926324] env[61987]: DEBUG nova.compute.manager [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Instance network_info: |[{"id": "4a1063a6-ed46-438e-af9b-6345483b1c9a", "address": "fa:16:3e:62:e1:24", "network": {"id": "ca68bf96-3f18-45ee-80fb-64a1260ed996", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-733595377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd47eb44bc334bf3ae5813905903ecbc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a27fd90b-16a5-43af-bede-ae36762ece00", "external-id": "nsx-vlan-transportzone-197", "segmentation_id": 197, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a1063a6-ed", "ovs_interfaceid": "4a1063a6-ed46-438e-af9b-6345483b1c9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 884.926914] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:e1:24', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a27fd90b-16a5-43af-bede-ae36762ece00', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4a1063a6-ed46-438e-af9b-6345483b1c9a', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 884.936177] env[61987]: DEBUG oslo.service.loopingcall [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.936825] env[61987]: DEBUG nova.compute.manager [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 884.937044] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 884.937898] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 884.938715] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f825c42-aecf-4ed7-ad33-06781e52ff17 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.943550] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-97a9716d-e37e-45af-a867-f44ecb5d1f53 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.966444] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 884.966740] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9704a9d5-8864-4c11-8ef5-8514059451ac {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.969440] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 884.969440] env[61987]: value = "task-1061874" [ 884.969440] env[61987]: _type = "Task" [ 884.969440] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.976930] env[61987]: DEBUG oslo_vmware.api [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 884.976930] env[61987]: value = "task-1061875" [ 884.976930] env[61987]: _type = "Task" [ 884.976930] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.987518] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061874, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.993927] env[61987]: DEBUG oslo_vmware.api [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061875, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.110328] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d22dd68f-c248-4409-acb7-1aa45a6131d1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.122590] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-450c4419-425e-423e-a1d0-2e7a92478857 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.164480] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cd10083-c7fa-4546-b06f-b6950b2e250c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.175101] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab37c3e3-4256-47fb-baf0-0a45b859dd2f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.195069] env[61987]: DEBUG nova.compute.provider_tree [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.274795] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061873, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.398158] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "ad40937f-d5a9-4708-8b81-06087d38d765" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.398524] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ad40937f-d5a9-4708-8b81-06087d38d765" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.398761] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "ad40937f-d5a9-4708-8b81-06087d38d765-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.399127] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ad40937f-d5a9-4708-8b81-06087d38d765-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.399377] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ad40937f-d5a9-4708-8b81-06087d38d765-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.401943] env[61987]: INFO nova.compute.manager [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Terminating instance [ 885.483139] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061874, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.491926] env[61987]: DEBUG oslo_vmware.api [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061875, 'name': PowerOffVM_Task, 'duration_secs': 0.241552} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.493363] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 885.493625] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 885.494960] env[61987]: DEBUG nova.compute.manager [req-36098459-9263-4c23-acc1-eaa9a1371a1a req-e2786705-7f0d-4df4-8f52-dd07a0edaab5 service nova] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Received event network-changed-4a1063a6-ed46-438e-af9b-6345483b1c9a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 885.495173] env[61987]: DEBUG nova.compute.manager [req-36098459-9263-4c23-acc1-eaa9a1371a1a req-e2786705-7f0d-4df4-8f52-dd07a0edaab5 service nova] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Refreshing instance network info cache due to event network-changed-4a1063a6-ed46-438e-af9b-6345483b1c9a. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 885.495941] env[61987]: DEBUG oslo_concurrency.lockutils [req-36098459-9263-4c23-acc1-eaa9a1371a1a req-e2786705-7f0d-4df4-8f52-dd07a0edaab5 service nova] Acquiring lock "refresh_cache-0422be0a-213b-41e9-b850-c823b52a90ca" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.495941] env[61987]: DEBUG oslo_concurrency.lockutils [req-36098459-9263-4c23-acc1-eaa9a1371a1a req-e2786705-7f0d-4df4-8f52-dd07a0edaab5 service nova] Acquired lock "refresh_cache-0422be0a-213b-41e9-b850-c823b52a90ca" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.495941] env[61987]: DEBUG nova.network.neutron [req-36098459-9263-4c23-acc1-eaa9a1371a1a req-e2786705-7f0d-4df4-8f52-dd07a0edaab5 service nova] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Refreshing network info cache for port 4a1063a6-ed46-438e-af9b-6345483b1c9a {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 885.497104] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7f62a9ec-8cf6-4f8b-8f12-a1fabefc14fb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.572457] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 885.572821] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 885.573117] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Deleting the datastore file [datastore1] 742297eb-c45d-4f45-85d2-e9bdd3106d1e {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 885.573468] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b8c6de4d-682f-4e2d-9050-6689070dea83 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.582837] env[61987]: DEBUG oslo_vmware.api [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 885.582837] env[61987]: value = "task-1061877" [ 885.582837] env[61987]: _type = "Task" [ 885.582837] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.592590] env[61987]: DEBUG oslo_vmware.api [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061877, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.593861] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8d96470-8adc-4093-b2a3-f7b6ae6128cb tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "refresh_cache-ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.594145] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8d96470-8adc-4093-b2a3-f7b6ae6128cb tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired lock "refresh_cache-ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.594429] env[61987]: DEBUG nova.network.neutron [None req-b8d96470-8adc-4093-b2a3-f7b6ae6128cb tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 885.594713] env[61987]: DEBUG nova.objects.instance [None req-b8d96470-8adc-4093-b2a3-f7b6ae6128cb tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lazy-loading 'info_cache' on Instance uuid ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.698762] env[61987]: DEBUG nova.scheduler.client.report [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 885.778295] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061873, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.674467} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.778610] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 212e4860-751e-43f7-80c6-25b6784e3541/212e4860-751e-43f7-80c6-25b6784e3541.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 885.781009] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 885.781009] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-503af14d-e58d-4e5e-a4a8-bd510ea442a6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.792351] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 885.792351] env[61987]: value = "task-1061878" [ 885.792351] env[61987]: _type = "Task" [ 885.792351] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.800472] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061878, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.906949] env[61987]: DEBUG nova.compute.manager [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 885.907255] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 885.908176] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36422b06-1f22-4f78-9e12-866b3f2417ac {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.918585] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 885.918585] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-220e1454-f6e5-452a-b441-2362c0778ec2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.927904] env[61987]: DEBUG oslo_vmware.api [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 885.927904] env[61987]: value = "task-1061879" [ 885.927904] env[61987]: _type = "Task" [ 885.927904] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.939329] env[61987]: DEBUG oslo_vmware.api [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061879, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.988248] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061874, 'name': CreateVM_Task, 'duration_secs': 0.567519} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.988399] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 885.989141] env[61987]: DEBUG oslo_concurrency.lockutils [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.989320] env[61987]: DEBUG oslo_concurrency.lockutils [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.989672] env[61987]: DEBUG oslo_concurrency.lockutils [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 885.989951] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-581630d0-0d0c-4278-a32b-d171ea9962d9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.997288] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 885.997288] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]525c9daf-ad62-71bd-5d5e-11f94227875f" [ 885.997288] env[61987]: _type = "Task" [ 885.997288] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.015609] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]525c9daf-ad62-71bd-5d5e-11f94227875f, 'name': SearchDatastore_Task, 'duration_secs': 0.013323} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.015609] env[61987]: DEBUG oslo_concurrency.lockutils [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.015609] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 886.015925] env[61987]: DEBUG oslo_concurrency.lockutils [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.016410] env[61987]: DEBUG oslo_concurrency.lockutils [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.016803] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 886.017412] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca270d09-098a-43e5-b90a-41533f5b00c2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.027795] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 886.028519] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 886.028873] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e0481c6-7122-4e7a-b3bb-ffc1d49aade4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.041282] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 886.041282] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52a48822-9dfa-3508-4759-20042aadf34a" [ 886.041282] env[61987]: _type = "Task" [ 886.041282] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.053973] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52a48822-9dfa-3508-4759-20042aadf34a, 'name': SearchDatastore_Task, 'duration_secs': 0.012343} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.055214] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddead404-6a7e-4ad4-938e-0a66c2d50165 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.061684] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 886.061684] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]520d44f6-11d6-bd40-8cf3-c5e9c60de32c" [ 886.061684] env[61987]: _type = "Task" [ 886.061684] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.074425] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]520d44f6-11d6-bd40-8cf3-c5e9c60de32c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.097642] env[61987]: DEBUG oslo_vmware.api [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061877, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166248} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.098371] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 886.098718] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 886.099046] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 886.099552] env[61987]: INFO nova.compute.manager [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Took 1.16 seconds to destroy the instance on the hypervisor. [ 886.099947] env[61987]: DEBUG oslo.service.loopingcall [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 886.100475] env[61987]: DEBUG nova.compute.manager [-] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 886.100658] env[61987]: DEBUG nova.network.neutron [-] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 886.205371] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.907s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.208931] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.152s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.209803] env[61987]: DEBUG nova.objects.instance [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lazy-loading 'pci_requests' on Instance uuid 51a36102-795e-47b7-b96a-857e54dc703e {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 886.231548] env[61987]: INFO nova.scheduler.client.report [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Deleted allocations for instance 1f29d5c1-bd01-47dc-9dcf-4e89791120d0 [ 886.307514] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061878, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073487} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.307885] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 886.309760] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9593916e-d15d-4d54-94b7-a53f9efe1b72 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.337690] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 212e4860-751e-43f7-80c6-25b6784e3541/212e4860-751e-43f7-80c6-25b6784e3541.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 886.338261] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-daef922b-de50-4768-858a-d88f9b703efa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.365275] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 886.365275] env[61987]: value = "task-1061880" [ 886.365275] env[61987]: _type = "Task" [ 886.365275] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.379526] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061880, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.437931] env[61987]: DEBUG oslo_vmware.api [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061879, 'name': PowerOffVM_Task, 'duration_secs': 0.404133} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.438263] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 886.438460] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 886.438724] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3dcc751c-1217-4ce2-b859-1dc2115da4f2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.452628] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Acquiring lock "d11104e7-4a5c-44c3-bd92-d36c587da794" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.452937] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Lock "d11104e7-4a5c-44c3-bd92-d36c587da794" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.509775] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 886.510227] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 886.510559] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Deleting the datastore file [datastore2] ad40937f-d5a9-4708-8b81-06087d38d765 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 886.510963] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-463cf5fe-8f1a-4b60-80af-b91faee5dba8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.518254] env[61987]: DEBUG nova.network.neutron [req-36098459-9263-4c23-acc1-eaa9a1371a1a req-e2786705-7f0d-4df4-8f52-dd07a0edaab5 service nova] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Updated VIF entry in instance network info cache for port 4a1063a6-ed46-438e-af9b-6345483b1c9a. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 886.518807] env[61987]: DEBUG nova.network.neutron [req-36098459-9263-4c23-acc1-eaa9a1371a1a req-e2786705-7f0d-4df4-8f52-dd07a0edaab5 service nova] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Updating instance_info_cache with network_info: [{"id": "4a1063a6-ed46-438e-af9b-6345483b1c9a", "address": "fa:16:3e:62:e1:24", "network": {"id": "ca68bf96-3f18-45ee-80fb-64a1260ed996", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-733595377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd47eb44bc334bf3ae5813905903ecbc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a27fd90b-16a5-43af-bede-ae36762ece00", "external-id": "nsx-vlan-transportzone-197", "segmentation_id": 197, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4a1063a6-ed", "ovs_interfaceid": "4a1063a6-ed46-438e-af9b-6345483b1c9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.527503] env[61987]: DEBUG oslo_vmware.api [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 886.527503] env[61987]: value = "task-1061882" [ 886.527503] env[61987]: _type = "Task" [ 886.527503] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.540471] env[61987]: DEBUG oslo_vmware.api [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061882, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.576823] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]520d44f6-11d6-bd40-8cf3-c5e9c60de32c, 'name': SearchDatastore_Task, 'duration_secs': 0.012737} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.577317] env[61987]: DEBUG oslo_concurrency.lockutils [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.577722] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 0422be0a-213b-41e9-b850-c823b52a90ca/0422be0a-213b-41e9-b850-c823b52a90ca.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 886.578097] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cc0a05db-f186-4759-9336-7875dc7fa85a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.587578] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 886.587578] env[61987]: value = "task-1061883" [ 886.587578] env[61987]: _type = "Task" [ 886.587578] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.598250] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061883, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.715394] env[61987]: DEBUG nova.objects.instance [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lazy-loading 'numa_topology' on Instance uuid 51a36102-795e-47b7-b96a-857e54dc703e {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 886.744709] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dadd793a-193b-471a-809e-2d6e29375df9 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.635s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.749982] env[61987]: DEBUG oslo_concurrency.lockutils [req-cd01bbbd-ff95-47bc-8290-c1e6f29336c7 req-f8a134b1-ef1e-4a34-ada3-758f6ed6a7bc service nova] Acquired lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.749982] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8f29b2-a402-4d6c-8b82-3d2c9f7f5a7b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.757847] env[61987]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 886.758817] env[61987]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=61987) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 886.758995] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ba3f61be-f2d1-4ee9-8341-eeb7343383c2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.770597] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7390ca1a-aeeb-42fa-9a30-68f942007377 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.804616] env[61987]: ERROR root [req-cd01bbbd-ff95-47bc-8290-c1e6f29336c7 req-f8a134b1-ef1e-4a34-ada3-758f6ed6a7bc service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-234333' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-234333' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-234333' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-234333'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-234333' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-234333' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-234333'}\n"]: nova.exception.InstanceNotFound: Instance 1f29d5c1-bd01-47dc-9dcf-4e89791120d0 could not be found. [ 886.805142] env[61987]: DEBUG oslo_concurrency.lockutils [req-cd01bbbd-ff95-47bc-8290-c1e6f29336c7 req-f8a134b1-ef1e-4a34-ada3-758f6ed6a7bc service nova] Releasing lock "1f29d5c1-bd01-47dc-9dcf-4e89791120d0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.805320] env[61987]: DEBUG nova.compute.manager [req-cd01bbbd-ff95-47bc-8290-c1e6f29336c7 req-f8a134b1-ef1e-4a34-ada3-758f6ed6a7bc service nova] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Detach interface failed, port_id=ebd9819d-a92d-4034-a00b-af37dbb338ab, reason: Instance 1f29d5c1-bd01-47dc-9dcf-4e89791120d0 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 886.878440] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061880, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.955933] env[61987]: DEBUG nova.compute.manager [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 887.021981] env[61987]: DEBUG oslo_concurrency.lockutils [req-36098459-9263-4c23-acc1-eaa9a1371a1a req-e2786705-7f0d-4df4-8f52-dd07a0edaab5 service nova] Releasing lock "refresh_cache-0422be0a-213b-41e9-b850-c823b52a90ca" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.041349] env[61987]: DEBUG oslo_vmware.api [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061882, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.173631} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.041684] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 887.041885] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 887.042086] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 887.042278] env[61987]: INFO nova.compute.manager [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Took 1.14 seconds to destroy the instance on the hypervisor. [ 887.042547] env[61987]: DEBUG oslo.service.loopingcall [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 887.042784] env[61987]: DEBUG nova.compute.manager [-] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 887.042887] env[61987]: DEBUG nova.network.neutron [-] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 887.104738] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061883, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.218563] env[61987]: INFO nova.compute.claims [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 887.237372] env[61987]: DEBUG nova.network.neutron [None req-b8d96470-8adc-4093-b2a3-f7b6ae6128cb tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Updating instance_info_cache with network_info: [{"id": "507c557c-1a2c-4bcb-a8da-93fc1fa91bd9", "address": "fa:16:3e:8c:d5:4a", "network": {"id": "59d8d7d6-fd71-4dec-802e-e0f9e670ff70", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1845227031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "315a2a3450024ed38e27cd7758f392f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap507c557c-1a", "ovs_interfaceid": "507c557c-1a2c-4bcb-a8da-93fc1fa91bd9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.382022] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061880, 'name': ReconfigVM_Task, 'duration_secs': 0.546458} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.383754] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 212e4860-751e-43f7-80c6-25b6784e3541/212e4860-751e-43f7-80c6-25b6784e3541.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 887.384633] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3838eef1-35a4-4fd6-9f55-19ac0b6c2139 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.395636] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 887.395636] env[61987]: value = "task-1061884" [ 887.395636] env[61987]: _type = "Task" [ 887.395636] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.406383] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061884, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.513275] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.600740] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061883, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521739} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.601294] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 0422be0a-213b-41e9-b850-c823b52a90ca/0422be0a-213b-41e9-b850-c823b52a90ca.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 887.601456] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 887.601736] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b5c48ed3-d36c-4c93-9add-766fa7f810b3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.617837] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 887.617837] env[61987]: value = "task-1061885" [ 887.617837] env[61987]: _type = "Task" [ 887.617837] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.633374] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061885, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.705559] env[61987]: DEBUG nova.compute.manager [req-c1e96eac-dd27-47b0-9bc3-81dc53081699 req-446be4fc-689d-45d4-94af-d62657768c01 service nova] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Received event network-changed-785edefc-5e84-4a34-97e1-b7c1a08f218e {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 887.705774] env[61987]: DEBUG nova.compute.manager [req-c1e96eac-dd27-47b0-9bc3-81dc53081699 req-446be4fc-689d-45d4-94af-d62657768c01 service nova] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Refreshing instance network info cache due to event network-changed-785edefc-5e84-4a34-97e1-b7c1a08f218e. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 887.705999] env[61987]: DEBUG oslo_concurrency.lockutils [req-c1e96eac-dd27-47b0-9bc3-81dc53081699 req-446be4fc-689d-45d4-94af-d62657768c01 service nova] Acquiring lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.707030] env[61987]: DEBUG oslo_concurrency.lockutils [req-c1e96eac-dd27-47b0-9bc3-81dc53081699 req-446be4fc-689d-45d4-94af-d62657768c01 service nova] Acquired lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.707030] env[61987]: DEBUG nova.network.neutron [req-c1e96eac-dd27-47b0-9bc3-81dc53081699 req-446be4fc-689d-45d4-94af-d62657768c01 service nova] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Refreshing network info cache for port 785edefc-5e84-4a34-97e1-b7c1a08f218e {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 887.739881] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8d96470-8adc-4093-b2a3-f7b6ae6128cb tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Releasing lock "refresh_cache-ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.740198] env[61987]: DEBUG nova.objects.instance [None req-b8d96470-8adc-4093-b2a3-f7b6ae6128cb tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lazy-loading 'migration_context' on Instance uuid ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.886777] env[61987]: DEBUG nova.network.neutron [-] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.912222] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061884, 'name': Rename_Task, 'duration_secs': 0.185795} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.912222] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 887.912222] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-207c6f33-c566-4ba8-b4ae-94e1490bbe7b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.920624] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 887.920624] env[61987]: value = "task-1061886" [ 887.920624] env[61987]: _type = "Task" [ 887.920624] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.934228] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061886, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.136979] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061885, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.1708} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.137774] env[61987]: DEBUG nova.compute.manager [req-5e55828b-64e6-4bab-b568-7760ea4809b0 req-7c373c0b-cd7c-40ef-8e60-43fb4626d061 service nova] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Received event network-vif-deleted-6aa03b99-d58f-461e-b2a8-c4552c1d418a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 888.138497] env[61987]: INFO nova.compute.manager [req-5e55828b-64e6-4bab-b568-7760ea4809b0 req-7c373c0b-cd7c-40ef-8e60-43fb4626d061 service nova] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Neutron deleted interface 6aa03b99-d58f-461e-b2a8-c4552c1d418a; detaching it from the instance and deleting it from the info cache [ 888.138497] env[61987]: DEBUG nova.network.neutron [req-5e55828b-64e6-4bab-b568-7760ea4809b0 req-7c373c0b-cd7c-40ef-8e60-43fb4626d061 service nova] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.139926] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 888.141596] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-251be541-567d-4dc0-8ba7-7ffbca46541c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.170345] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 0422be0a-213b-41e9-b850-c823b52a90ca/0422be0a-213b-41e9-b850-c823b52a90ca.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 888.171287] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ced0131e-ca51-4fad-aaad-6dc94276d008 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.194026] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 888.194026] env[61987]: value = "task-1061887" [ 888.194026] env[61987]: _type = "Task" [ 888.194026] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.203076] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061887, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.245861] env[61987]: DEBUG nova.objects.base [None req-b8d96470-8adc-4093-b2a3-f7b6ae6128cb tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61987) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 888.250789] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69a0d3bb-4095-4015-9de8-7f09ebe01b22 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.275279] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0576280-0680-4fde-8aa0-e291d53fe07a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.281786] env[61987]: DEBUG oslo_vmware.api [None req-b8d96470-8adc-4093-b2a3-f7b6ae6128cb tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 888.281786] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52e55a30-f970-7448-643c-1186061ff00e" [ 888.281786] env[61987]: _type = "Task" [ 888.281786] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.291570] env[61987]: DEBUG oslo_vmware.api [None req-b8d96470-8adc-4093-b2a3-f7b6ae6128cb tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52e55a30-f970-7448-643c-1186061ff00e, 'name': SearchDatastore_Task, 'duration_secs': 0.007645} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.291734] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8d96470-8adc-4093-b2a3-f7b6ae6128cb tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.389497] env[61987]: INFO nova.compute.manager [-] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Took 2.29 seconds to deallocate network for instance. [ 888.434942] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061886, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.512812] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d7dc14-1581-44d5-a596-1d76ee6d5cde {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.520994] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eb5ea68-cb24-45f6-93bd-706f821244d8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.553627] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35e8538a-fc56-4254-9036-7d284acd89ff {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.563703] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8014bcd4-c181-4a8d-8ff4-0e2d2da9f926 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.580182] env[61987]: DEBUG nova.compute.provider_tree [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.600936] env[61987]: DEBUG nova.network.neutron [-] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.629155] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Didn't find any instances for network info cache update. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10487}} [ 888.629691] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 888.629957] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 888.630180] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 888.630377] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 888.630580] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 888.630788] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 888.630929] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61987) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11020}} [ 888.631092] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 888.642060] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d967ffcd-eb79-4251-9e90-b40aad83e55b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.653796] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b846c8d6-83f4-4f7b-8501-57f970a90158 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.688818] env[61987]: DEBUG nova.compute.manager [req-5e55828b-64e6-4bab-b568-7760ea4809b0 req-7c373c0b-cd7c-40ef-8e60-43fb4626d061 service nova] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Detach interface failed, port_id=6aa03b99-d58f-461e-b2a8-c4552c1d418a, reason: Instance ad40937f-d5a9-4708-8b81-06087d38d765 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 888.704146] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061887, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.772472] env[61987]: DEBUG nova.network.neutron [req-c1e96eac-dd27-47b0-9bc3-81dc53081699 req-446be4fc-689d-45d4-94af-d62657768c01 service nova] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updated VIF entry in instance network info cache for port 785edefc-5e84-4a34-97e1-b7c1a08f218e. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 888.772941] env[61987]: DEBUG nova.network.neutron [req-c1e96eac-dd27-47b0-9bc3-81dc53081699 req-446be4fc-689d-45d4-94af-d62657768c01 service nova] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updating instance_info_cache with network_info: [{"id": "785edefc-5e84-4a34-97e1-b7c1a08f218e", "address": "fa:16:3e:bf:8b:d7", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap785edefc-5e", "ovs_interfaceid": "785edefc-5e84-4a34-97e1-b7c1a08f218e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.898651] env[61987]: DEBUG oslo_concurrency.lockutils [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.935566] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061886, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.085731] env[61987]: DEBUG nova.scheduler.client.report [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 889.106853] env[61987]: INFO nova.compute.manager [-] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Took 2.06 seconds to deallocate network for instance. [ 889.137338] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.204815] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061887, 'name': ReconfigVM_Task, 'duration_secs': 0.76238} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.204815] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 0422be0a-213b-41e9-b850-c823b52a90ca/0422be0a-213b-41e9-b850-c823b52a90ca.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 889.205468] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0817d721-27da-40fe-b9c1-9dc0123822d1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.216896] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 889.216896] env[61987]: value = "task-1061888" [ 889.216896] env[61987]: _type = "Task" [ 889.216896] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.229754] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061888, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.275734] env[61987]: DEBUG oslo_concurrency.lockutils [req-c1e96eac-dd27-47b0-9bc3-81dc53081699 req-446be4fc-689d-45d4-94af-d62657768c01 service nova] Releasing lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.276239] env[61987]: DEBUG nova.compute.manager [req-c1e96eac-dd27-47b0-9bc3-81dc53081699 req-446be4fc-689d-45d4-94af-d62657768c01 service nova] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Received event network-vif-deleted-0a4d6904-241e-451e-86d4-2cec32cf3b25 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 889.276508] env[61987]: INFO nova.compute.manager [req-c1e96eac-dd27-47b0-9bc3-81dc53081699 req-446be4fc-689d-45d4-94af-d62657768c01 service nova] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Neutron deleted interface 0a4d6904-241e-451e-86d4-2cec32cf3b25; detaching it from the instance and deleting it from the info cache [ 889.276827] env[61987]: DEBUG nova.network.neutron [req-c1e96eac-dd27-47b0-9bc3-81dc53081699 req-446be4fc-689d-45d4-94af-d62657768c01 service nova] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.440164] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061886, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.595405] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.387s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.597875] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.147s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.598157] env[61987]: DEBUG nova.objects.instance [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Lazy-loading 'resources' on Instance uuid 93706a54-1fc4-4aed-8807-ec2c3d3694b4 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 889.612305] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.637230] env[61987]: INFO nova.network.neutron [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Updating port 4e18ece6-5812-4c5c-827d-b3ee7d399008 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 889.727581] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061888, 'name': Rename_Task, 'duration_secs': 0.237203} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.727922] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 889.728200] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3e11b9a5-c81c-462d-be8b-ccb2d51c0f56 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.736113] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 889.736113] env[61987]: value = "task-1061889" [ 889.736113] env[61987]: _type = "Task" [ 889.736113] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.745218] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061889, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.783169] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ef11e7a1-1a0f-424e-94b0-aaabe383ceca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.793061] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-187a69e5-f3d2-4eee-bf77-119aca6bba5e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.827245] env[61987]: DEBUG nova.compute.manager [req-c1e96eac-dd27-47b0-9bc3-81dc53081699 req-446be4fc-689d-45d4-94af-d62657768c01 service nova] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Detach interface failed, port_id=0a4d6904-241e-451e-86d4-2cec32cf3b25, reason: Instance 742297eb-c45d-4f45-85d2-e9bdd3106d1e could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 889.935748] env[61987]: DEBUG oslo_vmware.api [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061886, 'name': PowerOnVM_Task, 'duration_secs': 1.562223} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.936092] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 889.936288] env[61987]: INFO nova.compute.manager [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Took 11.62 seconds to spawn the instance on the hypervisor. [ 889.936534] env[61987]: DEBUG nova.compute.manager [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 889.937372] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04245845-8175-4749-a333-6c86d17ecefe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.159808] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.160099] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.248924] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061889, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.305283] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee3a209c-8469-4222-9f59-31b57197eb84 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.313698] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e738b841-eb24-4e25-9abb-e6dd9787aba9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.346071] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1fdd0ba-daf4-4499-9860-91f3879d1757 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.354179] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cae6bb6-e671-4f34-bba0-0d590552a449 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.367904] env[61987]: DEBUG nova.compute.provider_tree [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.456817] env[61987]: INFO nova.compute.manager [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Took 22.92 seconds to build instance. [ 890.662782] env[61987]: DEBUG nova.compute.manager [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 890.751806] env[61987]: DEBUG oslo_vmware.api [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061889, 'name': PowerOnVM_Task, 'duration_secs': 0.965642} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.752123] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 890.752337] env[61987]: INFO nova.compute.manager [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Took 8.17 seconds to spawn the instance on the hypervisor. [ 890.752524] env[61987]: DEBUG nova.compute.manager [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 890.753386] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-121f6fd5-a550-4be1-baa2-19a60dc8d4ed {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.871423] env[61987]: DEBUG nova.scheduler.client.report [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 890.959203] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1517126a-d53f-4453-9a19-80493c994084 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "212e4860-751e-43f7-80c6-25b6784e3541" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.427s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.112945] env[61987]: DEBUG nova.compute.manager [req-9e5dc74f-5382-4f42-b190-5860944b1f6c req-2b635982-7a6c-4deb-86a5-96d2cbcb551d service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Received event network-vif-plugged-4e18ece6-5812-4c5c-827d-b3ee7d399008 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 891.112945] env[61987]: DEBUG oslo_concurrency.lockutils [req-9e5dc74f-5382-4f42-b190-5860944b1f6c req-2b635982-7a6c-4deb-86a5-96d2cbcb551d service nova] Acquiring lock "51a36102-795e-47b7-b96a-857e54dc703e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.113177] env[61987]: DEBUG oslo_concurrency.lockutils [req-9e5dc74f-5382-4f42-b190-5860944b1f6c req-2b635982-7a6c-4deb-86a5-96d2cbcb551d service nova] Lock "51a36102-795e-47b7-b96a-857e54dc703e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.113340] env[61987]: DEBUG oslo_concurrency.lockutils [req-9e5dc74f-5382-4f42-b190-5860944b1f6c req-2b635982-7a6c-4deb-86a5-96d2cbcb551d service nova] Lock "51a36102-795e-47b7-b96a-857e54dc703e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.113523] env[61987]: DEBUG nova.compute.manager [req-9e5dc74f-5382-4f42-b190-5860944b1f6c req-2b635982-7a6c-4deb-86a5-96d2cbcb551d service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] No waiting events found dispatching network-vif-plugged-4e18ece6-5812-4c5c-827d-b3ee7d399008 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 891.113695] env[61987]: WARNING nova.compute.manager [req-9e5dc74f-5382-4f42-b190-5860944b1f6c req-2b635982-7a6c-4deb-86a5-96d2cbcb551d service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Received unexpected event network-vif-plugged-4e18ece6-5812-4c5c-827d-b3ee7d399008 for instance with vm_state shelved_offloaded and task_state spawning. [ 891.185536] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.267982] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "refresh_cache-51a36102-795e-47b7-b96a-857e54dc703e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.268249] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquired lock "refresh_cache-51a36102-795e-47b7-b96a-857e54dc703e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.268472] env[61987]: DEBUG nova.network.neutron [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 891.271306] env[61987]: INFO nova.compute.manager [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Took 22.49 seconds to build instance. [ 891.378627] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.780s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.381063] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.886s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.382664] env[61987]: INFO nova.compute.claims [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 891.400739] env[61987]: INFO nova.scheduler.client.report [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Deleted allocations for instance 93706a54-1fc4-4aed-8807-ec2c3d3694b4 [ 891.720312] env[61987]: DEBUG nova.compute.manager [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 891.721299] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea0626a-b3a5-4213-9044-93f8d7471cf7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.774306] env[61987]: DEBUG oslo_concurrency.lockutils [None req-59d8bb64-ac76-4b74-97cb-afbf0b33a9d3 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "0422be0a-213b-41e9-b850-c823b52a90ca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.024s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.841843] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f7b2cad6-32db-4648-9ec2-85f0afa50815 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "0422be0a-213b-41e9-b850-c823b52a90ca" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.842112] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f7b2cad6-32db-4648-9ec2-85f0afa50815 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "0422be0a-213b-41e9-b850-c823b52a90ca" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.842315] env[61987]: DEBUG nova.compute.manager [None req-f7b2cad6-32db-4648-9ec2-85f0afa50815 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 891.843196] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cbb5d03-fb91-46cc-8067-c0aa0bffc880 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.851109] env[61987]: DEBUG nova.compute.manager [None req-f7b2cad6-32db-4648-9ec2-85f0afa50815 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61987) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 891.851714] env[61987]: DEBUG nova.objects.instance [None req-f7b2cad6-32db-4648-9ec2-85f0afa50815 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lazy-loading 'flavor' on Instance uuid 0422be0a-213b-41e9-b850-c823b52a90ca {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 891.910861] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ef6a7d14-be72-4824-b7ed-561d0657cd7d tempest-ServersWithSpecificFlavorTestJSON-885512048 tempest-ServersWithSpecificFlavorTestJSON-885512048-project-member] Lock "93706a54-1fc4-4aed-8807-ec2c3d3694b4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.274s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.973293] env[61987]: DEBUG nova.network.neutron [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Updating instance_info_cache with network_info: [{"id": "4e18ece6-5812-4c5c-827d-b3ee7d399008", "address": "fa:16:3e:1a:e6:c9", "network": {"id": "abcd2857-4c5a-4546-a003-858ddb17b0e6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-124838733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3fd98395e57149c987f82585296751b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e18ece6-58", "ovs_interfaceid": "4e18ece6-5812-4c5c-827d-b3ee7d399008", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.233894] env[61987]: INFO nova.compute.manager [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] instance snapshotting [ 892.237040] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a7438c-fe16-4387-a9a6-d606b9c4d97b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.258376] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6c09592-01b3-4908-a7ea-a4c15264219c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.480020] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Releasing lock "refresh_cache-51a36102-795e-47b7-b96a-857e54dc703e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.516910] env[61987]: DEBUG nova.virt.hardware [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='87e10c8e99ae9a7420c67463ef85a74c',container_format='bare',created_at=2024-09-30T23:12:46Z,direct_url=,disk_format='vmdk',id=87a46110-ae87-47e9-975e-48f434a17e9a,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-1771756833-shelved',owner='3fd98395e57149c987f82585296751b2',properties=ImageMetaProps,protected=,size=31667712,status='active',tags=,updated_at=2024-09-30T23:13:02Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 892.518018] env[61987]: DEBUG nova.virt.hardware [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 892.518018] env[61987]: DEBUG nova.virt.hardware [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 892.518018] env[61987]: DEBUG nova.virt.hardware [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 892.518018] env[61987]: DEBUG nova.virt.hardware [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 892.518018] env[61987]: DEBUG nova.virt.hardware [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 892.518221] env[61987]: DEBUG nova.virt.hardware [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 892.518261] env[61987]: DEBUG nova.virt.hardware [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 892.518658] env[61987]: DEBUG nova.virt.hardware [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 892.518658] env[61987]: DEBUG nova.virt.hardware [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 892.518793] env[61987]: DEBUG nova.virt.hardware [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 892.519647] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbe0f344-015c-4acb-b7aa-bb9a626e192d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.532888] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d00a784-9d6c-4195-891b-b4c400de7193 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.552917] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1a:e6:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '170f3b82-5915-4e36-bce9-4664ebb6be5e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4e18ece6-5812-4c5c-827d-b3ee7d399008', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 892.560416] env[61987]: DEBUG oslo.service.loopingcall [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 892.563942] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 892.564414] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3561abc1-5bf6-4a88-9e52-c13d264196e6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.590074] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 892.590074] env[61987]: value = "task-1061890" [ 892.590074] env[61987]: _type = "Task" [ 892.590074] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.599030] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061890, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.646514] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16f26109-73f9-4b61-8ddd-c5bdd00e7e58 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.655041] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55550b74-7f2f-403d-9b3c-2b60dd25eb2a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.691449] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17430c63-f100-439b-bdbd-6d63a32212eb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.700277] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5218d24-1793-4f09-af8d-93e144a7435a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.716676] env[61987]: DEBUG nova.compute.provider_tree [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.770363] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Creating Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 892.770363] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-3bdc6602-d756-4aa5-90ec-39e831ed500e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.779804] env[61987]: DEBUG oslo_vmware.api [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 892.779804] env[61987]: value = "task-1061891" [ 892.779804] env[61987]: _type = "Task" [ 892.779804] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.790229] env[61987]: DEBUG oslo_vmware.api [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061891, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.861791] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7b2cad6-32db-4648-9ec2-85f0afa50815 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 892.862176] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b3618f9-eaa4-45ef-98c9-0b0c412518c3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.872391] env[61987]: DEBUG oslo_vmware.api [None req-f7b2cad6-32db-4648-9ec2-85f0afa50815 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 892.872391] env[61987]: value = "task-1061892" [ 892.872391] env[61987]: _type = "Task" [ 892.872391] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.882065] env[61987]: DEBUG oslo_vmware.api [None req-f7b2cad6-32db-4648-9ec2-85f0afa50815 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061892, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.101134] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061890, 'name': CreateVM_Task, 'duration_secs': 0.508503} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.101519] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 893.102095] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/87a46110-ae87-47e9-975e-48f434a17e9a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.102249] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquired lock "[datastore2] devstack-image-cache_base/87a46110-ae87-47e9-975e-48f434a17e9a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.102700] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/87a46110-ae87-47e9-975e-48f434a17e9a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 893.102983] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a08c7cc-1b92-498b-b10d-c9e95d10de1c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.109102] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 893.109102] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]528133f5-5840-cd79-d965-e97ce45b989c" [ 893.109102] env[61987]: _type = "Task" [ 893.109102] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.117886] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]528133f5-5840-cd79-d965-e97ce45b989c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.143887] env[61987]: DEBUG nova.compute.manager [req-c36883c6-6048-4efb-be3d-7af046703987 req-efc9b404-be41-4161-90c4-1a1ac4bfb615 service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Received event network-changed-4e18ece6-5812-4c5c-827d-b3ee7d399008 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 893.144117] env[61987]: DEBUG nova.compute.manager [req-c36883c6-6048-4efb-be3d-7af046703987 req-efc9b404-be41-4161-90c4-1a1ac4bfb615 service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Refreshing instance network info cache due to event network-changed-4e18ece6-5812-4c5c-827d-b3ee7d399008. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 893.144338] env[61987]: DEBUG oslo_concurrency.lockutils [req-c36883c6-6048-4efb-be3d-7af046703987 req-efc9b404-be41-4161-90c4-1a1ac4bfb615 service nova] Acquiring lock "refresh_cache-51a36102-795e-47b7-b96a-857e54dc703e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.144522] env[61987]: DEBUG oslo_concurrency.lockutils [req-c36883c6-6048-4efb-be3d-7af046703987 req-efc9b404-be41-4161-90c4-1a1ac4bfb615 service nova] Acquired lock "refresh_cache-51a36102-795e-47b7-b96a-857e54dc703e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.144717] env[61987]: DEBUG nova.network.neutron [req-c36883c6-6048-4efb-be3d-7af046703987 req-efc9b404-be41-4161-90c4-1a1ac4bfb615 service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Refreshing network info cache for port 4e18ece6-5812-4c5c-827d-b3ee7d399008 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 893.220303] env[61987]: DEBUG nova.scheduler.client.report [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 893.293841] env[61987]: DEBUG oslo_vmware.api [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061891, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.382639] env[61987]: DEBUG oslo_vmware.api [None req-f7b2cad6-32db-4648-9ec2-85f0afa50815 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061892, 'name': PowerOffVM_Task, 'duration_secs': 0.227308} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.382947] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7b2cad6-32db-4648-9ec2-85f0afa50815 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 893.383723] env[61987]: DEBUG nova.compute.manager [None req-f7b2cad6-32db-4648-9ec2-85f0afa50815 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 893.384348] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f29d5672-484e-4f9d-b0a7-8bf9fa25f803 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.621386] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Releasing lock "[datastore2] devstack-image-cache_base/87a46110-ae87-47e9-975e-48f434a17e9a" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.621745] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Processing image 87a46110-ae87-47e9-975e-48f434a17e9a {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 893.621955] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/87a46110-ae87-47e9-975e-48f434a17e9a/87a46110-ae87-47e9-975e-48f434a17e9a.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.622148] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquired lock "[datastore2] devstack-image-cache_base/87a46110-ae87-47e9-975e-48f434a17e9a/87a46110-ae87-47e9-975e-48f434a17e9a.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.622426] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 893.622740] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-33a76bf0-ce96-4934-881a-bfdce5174eb5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.632835] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 893.633043] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 893.633817] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f80181b1-5b6c-47a6-9ded-d2fafc48d693 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.640187] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 893.640187] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52bbedca-8615-fbfd-37e2-b3f7e3c8dda7" [ 893.640187] env[61987]: _type = "Task" [ 893.640187] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.650865] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52bbedca-8615-fbfd-37e2-b3f7e3c8dda7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.725074] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.344s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.725728] env[61987]: DEBUG nova.compute.manager [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 893.728346] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.304s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.728580] env[61987]: DEBUG nova.objects.instance [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lazy-loading 'resources' on Instance uuid 769ecb4d-8fdd-4993-8a29-84c30d1089b2 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.792213] env[61987]: DEBUG oslo_vmware.api [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061891, 'name': CreateSnapshot_Task, 'duration_secs': 0.868647} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.792213] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Created Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 893.792979] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d6a17fb-fb10-412e-b4c6-1baed11ed474 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.895227] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f7b2cad6-32db-4648-9ec2-85f0afa50815 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "0422be0a-213b-41e9-b850-c823b52a90ca" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.053s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.931264] env[61987]: DEBUG nova.network.neutron [req-c36883c6-6048-4efb-be3d-7af046703987 req-efc9b404-be41-4161-90c4-1a1ac4bfb615 service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Updated VIF entry in instance network info cache for port 4e18ece6-5812-4c5c-827d-b3ee7d399008. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 893.931659] env[61987]: DEBUG nova.network.neutron [req-c36883c6-6048-4efb-be3d-7af046703987 req-efc9b404-be41-4161-90c4-1a1ac4bfb615 service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Updating instance_info_cache with network_info: [{"id": "4e18ece6-5812-4c5c-827d-b3ee7d399008", "address": "fa:16:3e:1a:e6:c9", "network": {"id": "abcd2857-4c5a-4546-a003-858ddb17b0e6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-124838733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3fd98395e57149c987f82585296751b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e18ece6-58", "ovs_interfaceid": "4e18ece6-5812-4c5c-827d-b3ee7d399008", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.151425] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Preparing fetch location {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 894.153243] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Fetch image to [datastore2] OSTACK_IMG_194acccb-ecfd-49b1-a765-81c54a1a1a6b/OSTACK_IMG_194acccb-ecfd-49b1-a765-81c54a1a1a6b.vmdk {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 894.153243] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Downloading stream optimized image 87a46110-ae87-47e9-975e-48f434a17e9a to [datastore2] OSTACK_IMG_194acccb-ecfd-49b1-a765-81c54a1a1a6b/OSTACK_IMG_194acccb-ecfd-49b1-a765-81c54a1a1a6b.vmdk on the data store datastore2 as vApp {{(pid=61987) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 894.153243] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Downloading image file data 87a46110-ae87-47e9-975e-48f434a17e9a to the ESX as VM named 'OSTACK_IMG_194acccb-ecfd-49b1-a765-81c54a1a1a6b' {{(pid=61987) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 894.231328] env[61987]: DEBUG nova.compute.utils [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 894.236689] env[61987]: DEBUG nova.compute.manager [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 894.236989] env[61987]: DEBUG nova.network.neutron [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 894.245604] env[61987]: DEBUG oslo_vmware.rw_handles [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 894.245604] env[61987]: value = "resgroup-9" [ 894.245604] env[61987]: _type = "ResourcePool" [ 894.245604] env[61987]: }. {{(pid=61987) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 894.245604] env[61987]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-9cac1046-67ff-4990-985d-aab053d9f39a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.273564] env[61987]: DEBUG oslo_vmware.rw_handles [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lease: (returnval){ [ 894.273564] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52801856-3766-5dec-ed38-e69372475eb4" [ 894.273564] env[61987]: _type = "HttpNfcLease" [ 894.273564] env[61987]: } obtained for vApp import into resource pool (val){ [ 894.273564] env[61987]: value = "resgroup-9" [ 894.273564] env[61987]: _type = "ResourcePool" [ 894.273564] env[61987]: }. {{(pid=61987) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 894.273847] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the lease: (returnval){ [ 894.273847] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52801856-3766-5dec-ed38-e69372475eb4" [ 894.273847] env[61987]: _type = "HttpNfcLease" [ 894.273847] env[61987]: } to be ready. {{(pid=61987) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 894.283568] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 894.283568] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52801856-3766-5dec-ed38-e69372475eb4" [ 894.283568] env[61987]: _type = "HttpNfcLease" [ 894.283568] env[61987]: } is initializing. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 894.286262] env[61987]: DEBUG nova.policy [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '102f79eb06304aebbe40ff6ae9f89376', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e43b1d992c07429685f7c89f1a8cfc9a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 894.317221] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Creating linked-clone VM from snapshot {{(pid=61987) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 894.318630] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0e646216-c87e-4de3-badf-52e49f03fe34 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.326367] env[61987]: DEBUG oslo_concurrency.lockutils [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "0422be0a-213b-41e9-b850-c823b52a90ca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.326622] env[61987]: DEBUG oslo_concurrency.lockutils [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "0422be0a-213b-41e9-b850-c823b52a90ca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.326905] env[61987]: DEBUG oslo_concurrency.lockutils [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "0422be0a-213b-41e9-b850-c823b52a90ca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.327075] env[61987]: DEBUG oslo_concurrency.lockutils [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "0422be0a-213b-41e9-b850-c823b52a90ca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.327495] env[61987]: DEBUG oslo_concurrency.lockutils [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "0422be0a-213b-41e9-b850-c823b52a90ca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.329461] env[61987]: INFO nova.compute.manager [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Terminating instance [ 894.335580] env[61987]: DEBUG oslo_vmware.api [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 894.335580] env[61987]: value = "task-1061894" [ 894.335580] env[61987]: _type = "Task" [ 894.335580] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.347575] env[61987]: DEBUG oslo_vmware.api [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061894, 'name': CloneVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.434651] env[61987]: DEBUG oslo_concurrency.lockutils [req-c36883c6-6048-4efb-be3d-7af046703987 req-efc9b404-be41-4161-90c4-1a1ac4bfb615 service nova] Releasing lock "refresh_cache-51a36102-795e-47b7-b96a-857e54dc703e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.504636] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e2d18e4-8a83-47d2-825e-a9cd082e3a0c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.513124] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f167d200-22ed-40b5-aa5d-d64e9bbe8078 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.545158] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ecb5e15-af33-4ed6-98fc-6190279dece4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.553853] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cb64d5d-af45-4586-921c-ad26dacfb62c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.570218] env[61987]: DEBUG nova.compute.provider_tree [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.602847] env[61987]: DEBUG nova.network.neutron [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Successfully created port: e81b029c-eef1-4ba6-8b98-a5647a44afe0 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 894.737736] env[61987]: DEBUG nova.compute.manager [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 894.783724] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 894.783724] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52801856-3766-5dec-ed38-e69372475eb4" [ 894.783724] env[61987]: _type = "HttpNfcLease" [ 894.783724] env[61987]: } is initializing. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 894.847900] env[61987]: DEBUG oslo_vmware.api [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061894, 'name': CloneVM_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.849496] env[61987]: DEBUG nova.compute.manager [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 894.849496] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 894.850144] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93274eeb-9172-48cc-b402-61916fff57b1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.859189] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 894.859471] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dc07149f-aa52-412d-9842-526d5f2877a6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.930013] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 894.930454] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 894.930454] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Deleting the datastore file [datastore2] 0422be0a-213b-41e9-b850-c823b52a90ca {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 894.930965] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f294efd7-b709-4d37-9c81-67cd9d563af4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.940925] env[61987]: DEBUG oslo_vmware.api [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 894.940925] env[61987]: value = "task-1061896" [ 894.940925] env[61987]: _type = "Task" [ 894.940925] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.949232] env[61987]: DEBUG oslo_vmware.api [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061896, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.074576] env[61987]: DEBUG nova.scheduler.client.report [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 895.285217] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 895.285217] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52801856-3766-5dec-ed38-e69372475eb4" [ 895.285217] env[61987]: _type = "HttpNfcLease" [ 895.285217] env[61987]: } is ready. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 895.286601] env[61987]: DEBUG oslo_vmware.rw_handles [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 895.286601] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52801856-3766-5dec-ed38-e69372475eb4" [ 895.286601] env[61987]: _type = "HttpNfcLease" [ 895.286601] env[61987]: }. {{(pid=61987) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 895.287739] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44b10da5-359d-494d-beae-b0ac2a85abd5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.297908] env[61987]: DEBUG oslo_vmware.rw_handles [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522ee514-79cd-04de-8660-4970d6851ab1/disk-0.vmdk from lease info. {{(pid=61987) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 895.298120] env[61987]: DEBUG oslo_vmware.rw_handles [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Creating HTTP connection to write to file with size = 31667712 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522ee514-79cd-04de-8660-4970d6851ab1/disk-0.vmdk. {{(pid=61987) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 895.369609] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-4d0c1eb8-d5e5-4627-b825-2d680c15b224 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.371659] env[61987]: DEBUG oslo_vmware.api [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061894, 'name': CloneVM_Task, 'duration_secs': 1.011634} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.373138] env[61987]: INFO nova.virt.vmwareapi.vmops [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Created linked-clone VM from snapshot [ 895.374221] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed91ec0-63d6-4827-8612-e29cfec7ed1e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.382964] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Uploading image c7aca900-ed48-48bb-a8f5-694fca140a87 {{(pid=61987) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 895.395377] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Destroying the VM {{(pid=61987) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 895.395586] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-1cc43fde-c793-4018-befa-d2beb4bd9c9a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.402389] env[61987]: DEBUG oslo_vmware.api [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 895.402389] env[61987]: value = "task-1061897" [ 895.402389] env[61987]: _type = "Task" [ 895.402389] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.410512] env[61987]: DEBUG oslo_vmware.api [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061897, 'name': Destroy_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.448892] env[61987]: DEBUG oslo_vmware.api [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061896, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.189884} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.449206] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 895.449441] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 895.449672] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 895.449890] env[61987]: INFO nova.compute.manager [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Took 0.60 seconds to destroy the instance on the hypervisor. [ 895.450183] env[61987]: DEBUG oslo.service.loopingcall [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.450410] env[61987]: DEBUG nova.compute.manager [-] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 895.450537] env[61987]: DEBUG nova.network.neutron [-] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 895.584968] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.856s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.587437] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.075s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.590549] env[61987]: INFO nova.compute.claims [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 895.639028] env[61987]: INFO nova.scheduler.client.report [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Deleted allocations for instance 769ecb4d-8fdd-4993-8a29-84c30d1089b2 [ 895.745177] env[61987]: DEBUG nova.compute.manager [req-d1f99601-4d89-447a-a4f3-80c2ca13a64e req-6bb9ccc8-9c7a-44e5-8f0c-1001fc6969ac service nova] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Received event network-vif-deleted-4a1063a6-ed46-438e-af9b-6345483b1c9a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 895.745288] env[61987]: INFO nova.compute.manager [req-d1f99601-4d89-447a-a4f3-80c2ca13a64e req-6bb9ccc8-9c7a-44e5-8f0c-1001fc6969ac service nova] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Neutron deleted interface 4a1063a6-ed46-438e-af9b-6345483b1c9a; detaching it from the instance and deleting it from the info cache [ 895.745472] env[61987]: DEBUG nova.network.neutron [req-d1f99601-4d89-447a-a4f3-80c2ca13a64e req-6bb9ccc8-9c7a-44e5-8f0c-1001fc6969ac service nova] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.754705] env[61987]: DEBUG nova.compute.manager [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 895.791921] env[61987]: DEBUG nova.virt.hardware [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 895.792212] env[61987]: DEBUG nova.virt.hardware [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 895.792458] env[61987]: DEBUG nova.virt.hardware [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 895.792721] env[61987]: DEBUG nova.virt.hardware [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 895.792885] env[61987]: DEBUG nova.virt.hardware [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 895.793102] env[61987]: DEBUG nova.virt.hardware [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 895.793311] env[61987]: DEBUG nova.virt.hardware [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 895.793540] env[61987]: DEBUG nova.virt.hardware [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 895.793768] env[61987]: DEBUG nova.virt.hardware [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 895.793988] env[61987]: DEBUG nova.virt.hardware [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 895.794408] env[61987]: DEBUG nova.virt.hardware [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 895.795696] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ebd65d-5108-4b69-9dcf-0961c926d687 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.809807] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75e24db6-496e-46af-b852-dbdb03d9ce22 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.919149] env[61987]: DEBUG oslo_vmware.api [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061897, 'name': Destroy_Task, 'duration_secs': 0.361079} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.921986] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Destroyed the VM [ 895.922381] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Deleting Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 895.922806] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e11de926-15d2-465f-8c9f-2ef6d8750286 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.931099] env[61987]: DEBUG oslo_vmware.api [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 895.931099] env[61987]: value = "task-1061898" [ 895.931099] env[61987]: _type = "Task" [ 895.931099] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.941960] env[61987]: DEBUG oslo_vmware.api [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061898, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.150059] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2c3ad168-f975-4a27-bf0d-077a3e98225e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "769ecb4d-8fdd-4993-8a29-84c30d1089b2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.314s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.220332] env[61987]: DEBUG nova.network.neutron [-] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.249385] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8ceff06b-7365-4549-a86b-ff177c70ef6a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.264457] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7869d4e-e077-4b24-bef5-0f47ebdab4c6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.283362] env[61987]: DEBUG nova.compute.manager [req-3598f17c-4305-40b6-b9aa-79223e2bb89c req-e798ae54-42fb-4fae-b350-d0ed413b5854 service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Received event network-vif-plugged-e81b029c-eef1-4ba6-8b98-a5647a44afe0 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 896.283654] env[61987]: DEBUG oslo_concurrency.lockutils [req-3598f17c-4305-40b6-b9aa-79223e2bb89c req-e798ae54-42fb-4fae-b350-d0ed413b5854 service nova] Acquiring lock "d8b9e172-c3bf-47d9-969b-7f33475d2d60-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.283875] env[61987]: DEBUG oslo_concurrency.lockutils [req-3598f17c-4305-40b6-b9aa-79223e2bb89c req-e798ae54-42fb-4fae-b350-d0ed413b5854 service nova] Lock "d8b9e172-c3bf-47d9-969b-7f33475d2d60-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.284193] env[61987]: DEBUG oslo_concurrency.lockutils [req-3598f17c-4305-40b6-b9aa-79223e2bb89c req-e798ae54-42fb-4fae-b350-d0ed413b5854 service nova] Lock "d8b9e172-c3bf-47d9-969b-7f33475d2d60-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.284362] env[61987]: DEBUG nova.compute.manager [req-3598f17c-4305-40b6-b9aa-79223e2bb89c req-e798ae54-42fb-4fae-b350-d0ed413b5854 service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] No waiting events found dispatching network-vif-plugged-e81b029c-eef1-4ba6-8b98-a5647a44afe0 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 896.284598] env[61987]: WARNING nova.compute.manager [req-3598f17c-4305-40b6-b9aa-79223e2bb89c req-e798ae54-42fb-4fae-b350-d0ed413b5854 service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Received unexpected event network-vif-plugged-e81b029c-eef1-4ba6-8b98-a5647a44afe0 for instance with vm_state building and task_state spawning. [ 896.303316] env[61987]: DEBUG nova.compute.manager [req-d1f99601-4d89-447a-a4f3-80c2ca13a64e req-6bb9ccc8-9c7a-44e5-8f0c-1001fc6969ac service nova] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Detach interface failed, port_id=4a1063a6-ed46-438e-af9b-6345483b1c9a, reason: Instance 0422be0a-213b-41e9-b850-c823b52a90ca could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 896.394217] env[61987]: DEBUG nova.network.neutron [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Successfully updated port: e81b029c-eef1-4ba6-8b98-a5647a44afe0 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 896.447230] env[61987]: DEBUG oslo_vmware.api [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061898, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.725581] env[61987]: INFO nova.compute.manager [-] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Took 1.27 seconds to deallocate network for instance. [ 896.772037] env[61987]: DEBUG oslo_vmware.rw_handles [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Completed reading data from the image iterator. {{(pid=61987) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 896.772037] env[61987]: DEBUG oslo_vmware.rw_handles [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522ee514-79cd-04de-8660-4970d6851ab1/disk-0.vmdk. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 896.774430] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90d2e629-9a5c-4bed-a0a4-0b7ff71d8688 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.780227] env[61987]: DEBUG oslo_vmware.rw_handles [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522ee514-79cd-04de-8660-4970d6851ab1/disk-0.vmdk is in state: ready. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 896.781036] env[61987]: DEBUG oslo_vmware.rw_handles [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522ee514-79cd-04de-8660-4970d6851ab1/disk-0.vmdk. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 896.781036] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-20505458-c4aa-4358-91c6-e408d23c3147 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.840821] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b9c6d8a-87b6-4d53-87e0-6b8ffaf26568 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.850614] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf640a2e-60da-4a07-baad-c285873efee9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.884220] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dde1a3f-8dc8-4967-9777-33b314715685 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.892587] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d75f414d-ac7c-4957-ac84-3e668c07a735 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.906720] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquiring lock "refresh_cache-d8b9e172-c3bf-47d9-969b-7f33475d2d60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.906878] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquired lock "refresh_cache-d8b9e172-c3bf-47d9-969b-7f33475d2d60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.907044] env[61987]: DEBUG nova.network.neutron [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 896.908390] env[61987]: DEBUG nova.compute.provider_tree [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.944336] env[61987]: DEBUG oslo_vmware.api [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061898, 'name': RemoveSnapshot_Task, 'duration_secs': 0.706662} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.944995] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Deleted Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 897.237190] env[61987]: DEBUG oslo_concurrency.lockutils [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.368648] env[61987]: DEBUG oslo_vmware.rw_handles [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522ee514-79cd-04de-8660-4970d6851ab1/disk-0.vmdk. {{(pid=61987) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 897.368965] env[61987]: INFO nova.virt.vmwareapi.images [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Downloaded image file data 87a46110-ae87-47e9-975e-48f434a17e9a [ 897.369864] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fbeaaa1-21c1-47a3-a9d8-683902c38fa6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.386993] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d2b04911-354a-4b6e-98d3-2d667407724d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.413683] env[61987]: DEBUG nova.scheduler.client.report [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 897.449388] env[61987]: WARNING nova.compute.manager [None req-d8904d7e-88ac-4900-8b67-4a169f7f0a91 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Image not found during snapshot: nova.exception.ImageNotFound: Image c7aca900-ed48-48bb-a8f5-694fca140a87 could not be found. [ 897.469827] env[61987]: DEBUG nova.network.neutron [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 897.498395] env[61987]: INFO nova.virt.vmwareapi.images [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] The imported VM was unregistered [ 897.500879] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Caching image {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 897.501333] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Creating directory with path [datastore2] devstack-image-cache_base/87a46110-ae87-47e9-975e-48f434a17e9a {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 897.501830] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-efcd9651-d98e-4c70-b04f-2573860eb7b3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.513206] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Created directory with path [datastore2] devstack-image-cache_base/87a46110-ae87-47e9-975e-48f434a17e9a {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 897.513562] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_194acccb-ecfd-49b1-a765-81c54a1a1a6b/OSTACK_IMG_194acccb-ecfd-49b1-a765-81c54a1a1a6b.vmdk to [datastore2] devstack-image-cache_base/87a46110-ae87-47e9-975e-48f434a17e9a/87a46110-ae87-47e9-975e-48f434a17e9a.vmdk. {{(pid=61987) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 897.514533] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-32c0b598-efbd-4556-b401-20eabcd99360 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.522358] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 897.522358] env[61987]: value = "task-1061900" [ 897.522358] env[61987]: _type = "Task" [ 897.522358] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.531270] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061900, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.700025] env[61987]: DEBUG nova.network.neutron [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Updating instance_info_cache with network_info: [{"id": "e81b029c-eef1-4ba6-8b98-a5647a44afe0", "address": "fa:16:3e:b0:72:9f", "network": {"id": "56353fe7-1959-4e8d-943c-a1ee1ffc860f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-992817851-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e43b1d992c07429685f7c89f1a8cfc9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape81b029c-ee", "ovs_interfaceid": "e81b029c-eef1-4ba6-8b98-a5647a44afe0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.822799] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "1b08af4a-8e3e-405e-8526-1e0a37b1f0cd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.823158] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "1b08af4a-8e3e-405e-8526-1e0a37b1f0cd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.919094] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.332s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.919742] env[61987]: DEBUG nova.compute.manager [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 897.925109] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8d96470-8adc-4093-b2a3-f7b6ae6128cb tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 9.633s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.032413] env[61987]: DEBUG oslo_concurrency.lockutils [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.032936] env[61987]: DEBUG oslo_concurrency.lockutils [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.041471] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061900, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.203389] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Releasing lock "refresh_cache-d8b9e172-c3bf-47d9-969b-7f33475d2d60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.203776] env[61987]: DEBUG nova.compute.manager [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Instance network_info: |[{"id": "e81b029c-eef1-4ba6-8b98-a5647a44afe0", "address": "fa:16:3e:b0:72:9f", "network": {"id": "56353fe7-1959-4e8d-943c-a1ee1ffc860f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-992817851-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e43b1d992c07429685f7c89f1a8cfc9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape81b029c-ee", "ovs_interfaceid": "e81b029c-eef1-4ba6-8b98-a5647a44afe0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 898.204526] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:72:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '93c5b7ce-4c84-40bc-884c-b2453e0eee69', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e81b029c-eef1-4ba6-8b98-a5647a44afe0', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 898.216685] env[61987]: DEBUG oslo.service.loopingcall [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 898.216685] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 898.216685] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1be37953-c4c2-472d-82ff-e845c032cea8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.242121] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 898.242121] env[61987]: value = "task-1061901" [ 898.242121] env[61987]: _type = "Task" [ 898.242121] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.254338] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061901, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.329023] env[61987]: DEBUG nova.compute.manager [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 898.333507] env[61987]: DEBUG nova.compute.manager [req-6897a2a9-7310-4853-b396-a0ec2340bf34 req-eda269ec-8de8-4d96-8836-d4cc137d2fb1 service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Received event network-changed-e81b029c-eef1-4ba6-8b98-a5647a44afe0 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 898.333507] env[61987]: DEBUG nova.compute.manager [req-6897a2a9-7310-4853-b396-a0ec2340bf34 req-eda269ec-8de8-4d96-8836-d4cc137d2fb1 service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Refreshing instance network info cache due to event network-changed-e81b029c-eef1-4ba6-8b98-a5647a44afe0. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 898.333507] env[61987]: DEBUG oslo_concurrency.lockutils [req-6897a2a9-7310-4853-b396-a0ec2340bf34 req-eda269ec-8de8-4d96-8836-d4cc137d2fb1 service nova] Acquiring lock "refresh_cache-d8b9e172-c3bf-47d9-969b-7f33475d2d60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.333507] env[61987]: DEBUG oslo_concurrency.lockutils [req-6897a2a9-7310-4853-b396-a0ec2340bf34 req-eda269ec-8de8-4d96-8836-d4cc137d2fb1 service nova] Acquired lock "refresh_cache-d8b9e172-c3bf-47d9-969b-7f33475d2d60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.333730] env[61987]: DEBUG nova.network.neutron [req-6897a2a9-7310-4853-b396-a0ec2340bf34 req-eda269ec-8de8-4d96-8836-d4cc137d2fb1 service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Refreshing network info cache for port e81b029c-eef1-4ba6-8b98-a5647a44afe0 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 898.376879] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "286f24dd-2239-434e-8ae7-15164a0ea8b0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.377201] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "286f24dd-2239-434e-8ae7-15164a0ea8b0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.429287] env[61987]: DEBUG nova.compute.utils [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 898.436411] env[61987]: DEBUG nova.compute.manager [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 898.436605] env[61987]: DEBUG nova.network.neutron [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 898.510722] env[61987]: DEBUG nova.policy [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '479e41ef9c9e4b35b696192e80c2da0b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e55e55f3d692423298be14eec1fd2508', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 898.537784] env[61987]: DEBUG nova.compute.manager [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 898.543907] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061900, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.695299] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "212e4860-751e-43f7-80c6-25b6784e3541" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.695466] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "212e4860-751e-43f7-80c6-25b6784e3541" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.695730] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "212e4860-751e-43f7-80c6-25b6784e3541-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.695932] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "212e4860-751e-43f7-80c6-25b6784e3541-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.696591] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "212e4860-751e-43f7-80c6-25b6784e3541-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.700928] env[61987]: INFO nova.compute.manager [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Terminating instance [ 898.743755] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fab4510-5227-4213-bb63-aa73d67df4c8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.757427] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "876b74ae-cae8-44a9-96ba-e2869ed43be6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.759150] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "876b74ae-cae8-44a9-96ba-e2869ed43be6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.770642] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d1e4e3c-baa9-49ce-802f-157b436f83e2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.774059] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061901, 'name': CreateVM_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.804817] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9f2fd57-a633-4ae9-bc62-3dca42a493b9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.817021] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-030ee5af-f819-4fd4-b880-2600b62d5d11 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.830708] env[61987]: DEBUG nova.compute.provider_tree [None req-b8d96470-8adc-4093-b2a3-f7b6ae6128cb tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.854288] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.880541] env[61987]: DEBUG nova.compute.manager [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 898.937024] env[61987]: DEBUG nova.compute.manager [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 899.021014] env[61987]: DEBUG nova.network.neutron [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Successfully created port: 0e240bad-236d-411b-b538-f0025aa3c0ea {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 899.035613] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061900, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.071384] env[61987]: DEBUG oslo_concurrency.lockutils [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.151954] env[61987]: DEBUG nova.network.neutron [req-6897a2a9-7310-4853-b396-a0ec2340bf34 req-eda269ec-8de8-4d96-8836-d4cc137d2fb1 service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Updated VIF entry in instance network info cache for port e81b029c-eef1-4ba6-8b98-a5647a44afe0. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 899.152371] env[61987]: DEBUG nova.network.neutron [req-6897a2a9-7310-4853-b396-a0ec2340bf34 req-eda269ec-8de8-4d96-8836-d4cc137d2fb1 service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Updating instance_info_cache with network_info: [{"id": "e81b029c-eef1-4ba6-8b98-a5647a44afe0", "address": "fa:16:3e:b0:72:9f", "network": {"id": "56353fe7-1959-4e8d-943c-a1ee1ffc860f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-992817851-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e43b1d992c07429685f7c89f1a8cfc9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape81b029c-ee", "ovs_interfaceid": "e81b029c-eef1-4ba6-8b98-a5647a44afe0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.206216] env[61987]: DEBUG nova.compute.manager [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 899.206317] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 899.207308] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1627058-3f6a-41f9-b949-84d6b8212426 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.216608] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 899.216933] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6d9977c1-a0a9-4da3-8dab-064b69b3d001 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.224368] env[61987]: DEBUG oslo_vmware.api [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 899.224368] env[61987]: value = "task-1061902" [ 899.224368] env[61987]: _type = "Task" [ 899.224368] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.235815] env[61987]: DEBUG oslo_vmware.api [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061902, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.252592] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061901, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.265314] env[61987]: DEBUG nova.compute.manager [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 899.334774] env[61987]: DEBUG nova.scheduler.client.report [None req-b8d96470-8adc-4093-b2a3-f7b6ae6128cb tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 899.408591] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.539529] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061900, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.655813] env[61987]: DEBUG oslo_concurrency.lockutils [req-6897a2a9-7310-4853-b396-a0ec2340bf34 req-eda269ec-8de8-4d96-8836-d4cc137d2fb1 service nova] Releasing lock "refresh_cache-d8b9e172-c3bf-47d9-969b-7f33475d2d60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.736167] env[61987]: DEBUG oslo_vmware.api [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061902, 'name': PowerOffVM_Task, 'duration_secs': 0.319602} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.736504] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 899.736649] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 899.736939] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c3c6f763-8a90-4710-9dfe-abee794bbab0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.758111] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061901, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.805798] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 899.806068] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 899.809030] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Deleting the datastore file [datastore2] 212e4860-751e-43f7-80c6-25b6784e3541 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 899.809030] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.809030] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94014522-ccd7-48dd-a625-3e374cc0415d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.816390] env[61987]: DEBUG oslo_vmware.api [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for the task: (returnval){ [ 899.816390] env[61987]: value = "task-1061904" [ 899.816390] env[61987]: _type = "Task" [ 899.816390] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.825361] env[61987]: DEBUG oslo_vmware.api [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061904, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.946979] env[61987]: DEBUG nova.compute.manager [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 899.974659] env[61987]: DEBUG nova.virt.hardware [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 899.974935] env[61987]: DEBUG nova.virt.hardware [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 899.975114] env[61987]: DEBUG nova.virt.hardware [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 899.975311] env[61987]: DEBUG nova.virt.hardware [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 899.975464] env[61987]: DEBUG nova.virt.hardware [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 899.975619] env[61987]: DEBUG nova.virt.hardware [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 899.975837] env[61987]: DEBUG nova.virt.hardware [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 899.976010] env[61987]: DEBUG nova.virt.hardware [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 899.976191] env[61987]: DEBUG nova.virt.hardware [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 899.976430] env[61987]: DEBUG nova.virt.hardware [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 899.976538] env[61987]: DEBUG nova.virt.hardware [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 899.977557] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ef6bfbe-db97-4eae-bfbf-46e8f41fe110 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.986439] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c91b6450-aaec-493a-be45-b1e7f0224126 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.037020] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061900, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.511374} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.037336] env[61987]: INFO nova.virt.vmwareapi.ds_util [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_194acccb-ecfd-49b1-a765-81c54a1a1a6b/OSTACK_IMG_194acccb-ecfd-49b1-a765-81c54a1a1a6b.vmdk to [datastore2] devstack-image-cache_base/87a46110-ae87-47e9-975e-48f434a17e9a/87a46110-ae87-47e9-975e-48f434a17e9a.vmdk. [ 900.037563] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Cleaning up location [datastore2] OSTACK_IMG_194acccb-ecfd-49b1-a765-81c54a1a1a6b {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 900.037744] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_194acccb-ecfd-49b1-a765-81c54a1a1a6b {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 900.038018] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-11762f67-03f3-4ddd-9db4-53ac814979f2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.045336] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 900.045336] env[61987]: value = "task-1061905" [ 900.045336] env[61987]: _type = "Task" [ 900.045336] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.054054] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061905, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.255264] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061901, 'name': CreateVM_Task, 'duration_secs': 1.882237} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.255889] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 900.256237] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.256702] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.256702] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 900.257171] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf1aede1-3269-4352-aca7-37f8f1ea5a13 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.262166] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for the task: (returnval){ [ 900.262166] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52843e5e-0145-f853-5e79-61317b649311" [ 900.262166] env[61987]: _type = "Task" [ 900.262166] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.272436] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52843e5e-0145-f853-5e79-61317b649311, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.326217] env[61987]: DEBUG oslo_vmware.api [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Task: {'id': task-1061904, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.321932} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.326534] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 900.326742] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 900.326928] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 900.327127] env[61987]: INFO nova.compute.manager [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Took 1.12 seconds to destroy the instance on the hypervisor. [ 900.327382] env[61987]: DEBUG oslo.service.loopingcall [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.327587] env[61987]: DEBUG nova.compute.manager [-] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 900.327689] env[61987]: DEBUG nova.network.neutron [-] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 900.349273] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8d96470-8adc-4093-b2a3-f7b6ae6128cb tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.424s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.359018] env[61987]: DEBUG oslo_concurrency.lockutils [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.458s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.359018] env[61987]: DEBUG nova.objects.instance [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lazy-loading 'resources' on Instance uuid 742297eb-c45d-4f45-85d2-e9bdd3106d1e {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 900.566899] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061905, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.045538} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.567348] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 900.567629] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Releasing lock "[datastore2] devstack-image-cache_base/87a46110-ae87-47e9-975e-48f434a17e9a/87a46110-ae87-47e9-975e-48f434a17e9a.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.569097] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/87a46110-ae87-47e9-975e-48f434a17e9a/87a46110-ae87-47e9-975e-48f434a17e9a.vmdk to [datastore2] 51a36102-795e-47b7-b96a-857e54dc703e/51a36102-795e-47b7-b96a-857e54dc703e.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 900.569685] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7f315434-f85d-494d-9db2-4bf9f5b3d045 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.583309] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 900.583309] env[61987]: value = "task-1061906" [ 900.583309] env[61987]: _type = "Task" [ 900.583309] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.600841] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061906, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.689080] env[61987]: DEBUG nova.compute.manager [req-4c3e47cf-8f8c-40aa-9618-42b49c9ee0f3 req-8f2e472e-9336-4edf-9995-0b354fa00c67 service nova] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Received event network-vif-plugged-0e240bad-236d-411b-b538-f0025aa3c0ea {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 900.694815] env[61987]: DEBUG oslo_concurrency.lockutils [req-4c3e47cf-8f8c-40aa-9618-42b49c9ee0f3 req-8f2e472e-9336-4edf-9995-0b354fa00c67 service nova] Acquiring lock "d11104e7-4a5c-44c3-bd92-d36c587da794-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.695073] env[61987]: DEBUG oslo_concurrency.lockutils [req-4c3e47cf-8f8c-40aa-9618-42b49c9ee0f3 req-8f2e472e-9336-4edf-9995-0b354fa00c67 service nova] Lock "d11104e7-4a5c-44c3-bd92-d36c587da794-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.006s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.695239] env[61987]: DEBUG oslo_concurrency.lockutils [req-4c3e47cf-8f8c-40aa-9618-42b49c9ee0f3 req-8f2e472e-9336-4edf-9995-0b354fa00c67 service nova] Lock "d11104e7-4a5c-44c3-bd92-d36c587da794-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.695448] env[61987]: DEBUG nova.compute.manager [req-4c3e47cf-8f8c-40aa-9618-42b49c9ee0f3 req-8f2e472e-9336-4edf-9995-0b354fa00c67 service nova] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] No waiting events found dispatching network-vif-plugged-0e240bad-236d-411b-b538-f0025aa3c0ea {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 900.695691] env[61987]: WARNING nova.compute.manager [req-4c3e47cf-8f8c-40aa-9618-42b49c9ee0f3 req-8f2e472e-9336-4edf-9995-0b354fa00c67 service nova] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Received unexpected event network-vif-plugged-0e240bad-236d-411b-b538-f0025aa3c0ea for instance with vm_state building and task_state spawning. [ 900.744485] env[61987]: DEBUG nova.compute.manager [req-b8389ace-e40e-40a1-9471-aeead86f3476 req-926efa3f-46c7-4e45-a616-d72e22e32e59 service nova] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Received event network-vif-deleted-ee6aa50c-8e68-48dc-8986-87c78f51a88a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 900.744731] env[61987]: INFO nova.compute.manager [req-b8389ace-e40e-40a1-9471-aeead86f3476 req-926efa3f-46c7-4e45-a616-d72e22e32e59 service nova] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Neutron deleted interface ee6aa50c-8e68-48dc-8986-87c78f51a88a; detaching it from the instance and deleting it from the info cache [ 900.745015] env[61987]: DEBUG nova.network.neutron [req-b8389ace-e40e-40a1-9471-aeead86f3476 req-926efa3f-46c7-4e45-a616-d72e22e32e59 service nova] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.774712] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52843e5e-0145-f853-5e79-61317b649311, 'name': SearchDatastore_Task, 'duration_secs': 0.013548} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.774858] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.775044] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 900.775331] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.775515] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.775716] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 900.775991] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e8ccebea-c9fd-41a3-8abf-1d3d1045e51f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.795546] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 900.797667] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 900.797667] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-111adc1f-5cca-454a-bfe8-cd604c1ad85a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.803518] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for the task: (returnval){ [ 900.803518] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]529e5d17-ae24-f0d1-b4aa-577f4fca5f9b" [ 900.803518] env[61987]: _type = "Task" [ 900.803518] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.817695] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]529e5d17-ae24-f0d1-b4aa-577f4fca5f9b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.821164] env[61987]: DEBUG nova.network.neutron [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Successfully updated port: 0e240bad-236d-411b-b538-f0025aa3c0ea {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 900.916901] env[61987]: INFO nova.scheduler.client.report [None req-b8d96470-8adc-4093-b2a3-f7b6ae6128cb tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Deleted allocation for migration 2bef6405-895b-4208-9a03-bb4e85484aab [ 901.095108] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061906, 'name': CopyVirtualDisk_Task} progress is 18%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.121310] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c81d795-7a48-4c02-bb27-bd25477fee67 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.130281] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99254082-6f98-40e4-ab6e-50bfb42d54a4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.162543] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b32d95-43c7-4c39-8414-b69cf6abe139 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.171682] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba08c4e6-6818-480f-aae4-018745b938d1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.186552] env[61987]: DEBUG nova.compute.provider_tree [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 901.219455] env[61987]: DEBUG nova.network.neutron [-] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.247253] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2ff600ca-1a41-4400-9a3c-9708a64b4004 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.258965] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f004c1d3-7761-40b4-bf53-4c204f1e5de7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.291203] env[61987]: DEBUG nova.compute.manager [req-b8389ace-e40e-40a1-9471-aeead86f3476 req-926efa3f-46c7-4e45-a616-d72e22e32e59 service nova] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Detach interface failed, port_id=ee6aa50c-8e68-48dc-8986-87c78f51a88a, reason: Instance 212e4860-751e-43f7-80c6-25b6784e3541 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 901.315018] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]529e5d17-ae24-f0d1-b4aa-577f4fca5f9b, 'name': SearchDatastore_Task, 'duration_secs': 0.067563} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.317047] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ed7e275-5efe-46a4-a6be-2917149b9f77 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.323699] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Acquiring lock "refresh_cache-d11104e7-4a5c-44c3-bd92-d36c587da794" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.323904] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Acquired lock "refresh_cache-d11104e7-4a5c-44c3-bd92-d36c587da794" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.324053] env[61987]: DEBUG nova.network.neutron [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 901.325419] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for the task: (returnval){ [ 901.325419] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52421c7a-d023-f671-af8d-71cc4590b15f" [ 901.325419] env[61987]: _type = "Task" [ 901.325419] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.335720] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52421c7a-d023-f671-af8d-71cc4590b15f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.424867] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8d96470-8adc-4093-b2a3-f7b6ae6128cb tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 16.589s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.596448] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061906, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.690423] env[61987]: DEBUG nova.scheduler.client.report [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 901.721934] env[61987]: INFO nova.compute.manager [-] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Took 1.39 seconds to deallocate network for instance. [ 901.838362] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52421c7a-d023-f671-af8d-71cc4590b15f, 'name': SearchDatastore_Task, 'duration_secs': 0.072272} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.838506] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.838790] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] d8b9e172-c3bf-47d9-969b-7f33475d2d60/d8b9e172-c3bf-47d9-969b-7f33475d2d60.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 901.839694] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-61ccf3b2-654d-43e8-a439-191e3ed8335d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.847894] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for the task: (returnval){ [ 901.847894] env[61987]: value = "task-1061907" [ 901.847894] env[61987]: _type = "Task" [ 901.847894] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.858248] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061907, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.872967] env[61987]: DEBUG nova.network.neutron [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 901.995528] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.995841] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.996076] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.996266] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.996446] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.999830] env[61987]: INFO nova.compute.manager [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Terminating instance [ 902.075701] env[61987]: DEBUG nova.network.neutron [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Updating instance_info_cache with network_info: [{"id": "0e240bad-236d-411b-b538-f0025aa3c0ea", "address": "fa:16:3e:09:bd:1d", "network": {"id": "fa3476bc-abef-450e-9111-2e280a3f479a", "bridge": "br-int", "label": "tempest-ServersTestJSON-923853252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e55e55f3d692423298be14eec1fd2508", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "456bd8a2-0fb6-4b17-9d25-08e7995c5184", "external-id": "nsx-vlan-transportzone-65", "segmentation_id": 65, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e240bad-23", "ovs_interfaceid": "0e240bad-236d-411b-b538-f0025aa3c0ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.098141] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061906, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.196650] env[61987]: DEBUG oslo_concurrency.lockutils [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.840s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.200126] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 13.063s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.200876] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.200876] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61987) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 902.201228] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.589s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.201599] env[61987]: DEBUG nova.objects.instance [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lazy-loading 'resources' on Instance uuid ad40937f-d5a9-4708-8b81-06087d38d765 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.203819] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e447c6e-e87a-49d6-9375-ed2fb2f5095a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.216730] env[61987]: INFO nova.scheduler.client.report [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Deleted allocations for instance 742297eb-c45d-4f45-85d2-e9bdd3106d1e [ 902.219012] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3158650-9a14-4bec-9398-63ff3497a61c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.228799] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.246081] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c63eb915-c01f-4470-98a3-57e8f391beab {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.254983] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78f5abea-de8b-432b-ac49-0f1ccd1aa505 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.288741] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180562MB free_disk=179GB free_vcpus=48 pci_devices=None {{(pid=61987) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 902.288961] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.359261] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061907, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.511624] env[61987]: DEBUG nova.compute.manager [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 902.511941] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 902.513153] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-085e1d9b-5bc7-4e9d-aacd-042cef253161 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.522908] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 902.523310] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-efab478e-9d2c-443f-ab89-cca58fb58347 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.533653] env[61987]: DEBUG oslo_vmware.api [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 902.533653] env[61987]: value = "task-1061908" [ 902.533653] env[61987]: _type = "Task" [ 902.533653] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.544075] env[61987]: DEBUG oslo_vmware.api [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061908, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.579557] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Releasing lock "refresh_cache-d11104e7-4a5c-44c3-bd92-d36c587da794" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.580030] env[61987]: DEBUG nova.compute.manager [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Instance network_info: |[{"id": "0e240bad-236d-411b-b538-f0025aa3c0ea", "address": "fa:16:3e:09:bd:1d", "network": {"id": "fa3476bc-abef-450e-9111-2e280a3f479a", "bridge": "br-int", "label": "tempest-ServersTestJSON-923853252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e55e55f3d692423298be14eec1fd2508", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "456bd8a2-0fb6-4b17-9d25-08e7995c5184", "external-id": "nsx-vlan-transportzone-65", "segmentation_id": 65, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e240bad-23", "ovs_interfaceid": "0e240bad-236d-411b-b538-f0025aa3c0ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 902.580572] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:bd:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '456bd8a2-0fb6-4b17-9d25-08e7995c5184', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0e240bad-236d-411b-b538-f0025aa3c0ea', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 902.590097] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Creating folder: Project (e55e55f3d692423298be14eec1fd2508). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 902.590462] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-020601f8-5f6a-436f-a874-eafb0acdd8b9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.604980] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061906, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.610504] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Created folder: Project (e55e55f3d692423298be14eec1fd2508) in parent group-v234219. [ 902.610713] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Creating folder: Instances. Parent ref: group-v234377. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 902.611014] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-65fe5823-49b1-4c4f-b0cb-edf72005d1dc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.624812] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "ad9041d9-1e0f-46d6-93f7-efe82197deb6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.625103] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "ad9041d9-1e0f-46d6-93f7-efe82197deb6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.628695] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Created folder: Instances in parent group-v234377. [ 902.629011] env[61987]: DEBUG oslo.service.loopingcall [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 902.629461] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 902.629732] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7af311f5-d11a-4f4d-9480-27d68023e171 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.655591] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 902.655591] env[61987]: value = "task-1061911" [ 902.655591] env[61987]: _type = "Task" [ 902.655591] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.665093] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061911, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.725389] env[61987]: DEBUG nova.compute.manager [req-976d6322-6faa-461a-8530-47e9368577d9 req-727ac44b-ae98-4f51-8593-c7039db607bc service nova] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Received event network-changed-0e240bad-236d-411b-b538-f0025aa3c0ea {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 902.725597] env[61987]: DEBUG nova.compute.manager [req-976d6322-6faa-461a-8530-47e9368577d9 req-727ac44b-ae98-4f51-8593-c7039db607bc service nova] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Refreshing instance network info cache due to event network-changed-0e240bad-236d-411b-b538-f0025aa3c0ea. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 902.725821] env[61987]: DEBUG oslo_concurrency.lockutils [req-976d6322-6faa-461a-8530-47e9368577d9 req-727ac44b-ae98-4f51-8593-c7039db607bc service nova] Acquiring lock "refresh_cache-d11104e7-4a5c-44c3-bd92-d36c587da794" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.726940] env[61987]: DEBUG oslo_concurrency.lockutils [req-976d6322-6faa-461a-8530-47e9368577d9 req-727ac44b-ae98-4f51-8593-c7039db607bc service nova] Acquired lock "refresh_cache-d11104e7-4a5c-44c3-bd92-d36c587da794" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.726940] env[61987]: DEBUG nova.network.neutron [req-976d6322-6faa-461a-8530-47e9368577d9 req-727ac44b-ae98-4f51-8593-c7039db607bc service nova] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Refreshing network info cache for port 0e240bad-236d-411b-b538-f0025aa3c0ea {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 902.736365] env[61987]: DEBUG oslo_concurrency.lockutils [None req-12f01436-7259-4243-9c09-8f9bee30d535 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "742297eb-c45d-4f45-85d2-e9bdd3106d1e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.318s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.860582] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061907, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.028870] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8561cc33-e4e9-423e-b9c8-f355ffce8c00 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.041210] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e670dab8-def4-4e21-952e-e188f2d26459 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.047395] env[61987]: DEBUG oslo_vmware.api [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061908, 'name': PowerOffVM_Task, 'duration_secs': 0.322049} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.048098] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 903.048282] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 903.048565] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-964e8fb2-58b5-40e2-8ee0-f5c8451c594b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.080696] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f572a88e-821c-4864-9404-caccac26de4a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.090621] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c58c1bf0-7bf4-4678-bd7f-24b62d54f486 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.114907] env[61987]: DEBUG nova.compute.provider_tree [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.116995] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061906, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.49077} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.117316] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/87a46110-ae87-47e9-975e-48f434a17e9a/87a46110-ae87-47e9-975e-48f434a17e9a.vmdk to [datastore2] 51a36102-795e-47b7-b96a-857e54dc703e/51a36102-795e-47b7-b96a-857e54dc703e.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 903.118162] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76fb65b9-12cb-4aca-b95c-3eb043791340 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.133507] env[61987]: DEBUG nova.compute.manager [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 903.146724] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Reconfiguring VM instance instance-00000026 to attach disk [datastore2] 51a36102-795e-47b7-b96a-857e54dc703e/51a36102-795e-47b7-b96a-857e54dc703e.vmdk or device None with type streamOptimized {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 903.150659] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d420ef7b-c860-4009-81e7-1572bab0ce95 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.165169] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 903.165169] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 903.165169] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Deleting the datastore file [datastore2] ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 903.165473] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf427552-1655-466b-9c9c-39f0d5903a3b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.176411] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061911, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.178655] env[61987]: DEBUG oslo_vmware.api [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 903.178655] env[61987]: value = "task-1061914" [ 903.178655] env[61987]: _type = "Task" [ 903.178655] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.179034] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 903.179034] env[61987]: value = "task-1061913" [ 903.179034] env[61987]: _type = "Task" [ 903.179034] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.190381] env[61987]: DEBUG oslo_vmware.api [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061914, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.194448] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061913, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.360675] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061907, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.575401] env[61987]: DEBUG nova.network.neutron [req-976d6322-6faa-461a-8530-47e9368577d9 req-727ac44b-ae98-4f51-8593-c7039db607bc service nova] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Updated VIF entry in instance network info cache for port 0e240bad-236d-411b-b538-f0025aa3c0ea. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 903.575943] env[61987]: DEBUG nova.network.neutron [req-976d6322-6faa-461a-8530-47e9368577d9 req-727ac44b-ae98-4f51-8593-c7039db607bc service nova] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Updating instance_info_cache with network_info: [{"id": "0e240bad-236d-411b-b538-f0025aa3c0ea", "address": "fa:16:3e:09:bd:1d", "network": {"id": "fa3476bc-abef-450e-9111-2e280a3f479a", "bridge": "br-int", "label": "tempest-ServersTestJSON-923853252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e55e55f3d692423298be14eec1fd2508", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "456bd8a2-0fb6-4b17-9d25-08e7995c5184", "external-id": "nsx-vlan-transportzone-65", "segmentation_id": 65, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e240bad-23", "ovs_interfaceid": "0e240bad-236d-411b-b538-f0025aa3c0ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.617848] env[61987]: DEBUG nova.scheduler.client.report [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 903.665890] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.676272] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061911, 'name': CreateVM_Task, 'duration_secs': 0.59949} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.676446] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 903.677508] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.677699] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.678121] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 903.678311] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea12d83d-247e-4ede-8bf0-048a4e3ffa97 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.691141] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061913, 'name': ReconfigVM_Task, 'duration_secs': 0.456339} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.694596] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Reconfigured VM instance instance-00000026 to attach disk [datastore2] 51a36102-795e-47b7-b96a-857e54dc703e/51a36102-795e-47b7-b96a-857e54dc703e.vmdk or device None with type streamOptimized {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 903.695402] env[61987]: DEBUG oslo_vmware.api [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061914, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.472775} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.695677] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Waiting for the task: (returnval){ [ 903.695677] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52a981de-3c23-2d0a-6762-943c42a2b769" [ 903.695677] env[61987]: _type = "Task" [ 903.695677] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.695868] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-110b3d7e-9d5a-4c2a-a1cb-9eb34a87b858 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.697361] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 903.697557] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 903.697741] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 903.697922] env[61987]: INFO nova.compute.manager [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Took 1.19 seconds to destroy the instance on the hypervisor. [ 903.698234] env[61987]: DEBUG oslo.service.loopingcall [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 903.698504] env[61987]: DEBUG nova.compute.manager [-] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 903.698604] env[61987]: DEBUG nova.network.neutron [-] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 903.708569] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52a981de-3c23-2d0a-6762-943c42a2b769, 'name': SearchDatastore_Task, 'duration_secs': 0.009358} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.710518] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.710518] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 903.710518] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.710518] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.710734] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 903.711175] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 903.711175] env[61987]: value = "task-1061915" [ 903.711175] env[61987]: _type = "Task" [ 903.711175] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.711396] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ee78376c-6d9e-4089-9267-b4510af70832 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.722637] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061915, 'name': Rename_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.723709] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 903.723889] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 903.724621] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52d1d4be-a96c-40d3-a786-ed1bd46f37bf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.729449] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Waiting for the task: (returnval){ [ 903.729449] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]522af2f7-56ab-e38b-396f-97cece1f8496" [ 903.729449] env[61987]: _type = "Task" [ 903.729449] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.738669] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]522af2f7-56ab-e38b-396f-97cece1f8496, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.859366] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061907, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.735861} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.859696] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] d8b9e172-c3bf-47d9-969b-7f33475d2d60/d8b9e172-c3bf-47d9-969b-7f33475d2d60.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 903.859963] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 903.860283] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ad1ecdaf-16bd-4cab-9817-0c2c9c036152 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.866813] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for the task: (returnval){ [ 903.866813] env[61987]: value = "task-1061916" [ 903.866813] env[61987]: _type = "Task" [ 903.866813] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.875731] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061916, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.079747] env[61987]: DEBUG oslo_concurrency.lockutils [req-976d6322-6faa-461a-8530-47e9368577d9 req-727ac44b-ae98-4f51-8593-c7039db607bc service nova] Releasing lock "refresh_cache-d11104e7-4a5c-44c3-bd92-d36c587da794" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.124153] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.923s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.126630] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.941s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.128244] env[61987]: INFO nova.compute.claims [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 904.142497] env[61987]: INFO nova.scheduler.client.report [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Deleted allocations for instance ad40937f-d5a9-4708-8b81-06087d38d765 [ 904.222675] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061915, 'name': Rename_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.238656] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]522af2f7-56ab-e38b-396f-97cece1f8496, 'name': SearchDatastore_Task, 'duration_secs': 0.008545} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.239682] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d23ad777-1c2f-4281-a87b-3b6213a76e1a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.244340] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Waiting for the task: (returnval){ [ 904.244340] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]523cac0a-529b-55d5-3f88-c0aea079570f" [ 904.244340] env[61987]: _type = "Task" [ 904.244340] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.251912] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523cac0a-529b-55d5-3f88-c0aea079570f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.376960] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061916, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.259728} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.377196] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 904.377949] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1ef0374-a0ea-4023-9b0d-72b6b418d982 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.380455] env[61987]: DEBUG oslo_concurrency.lockutils [None req-40cec40b-9e09-4a21-b7de-85cda886d5b0 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "f937a5ec-a1d7-41d8-b998-fa18f545b304" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.380686] env[61987]: DEBUG oslo_concurrency.lockutils [None req-40cec40b-9e09-4a21-b7de-85cda886d5b0 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "f937a5ec-a1d7-41d8-b998-fa18f545b304" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.401475] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] d8b9e172-c3bf-47d9-969b-7f33475d2d60/d8b9e172-c3bf-47d9-969b-7f33475d2d60.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 904.402595] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87761476-c482-4ef5-b816-05c7aa8f62ae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.416614] env[61987]: DEBUG nova.network.neutron [-] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.423276] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for the task: (returnval){ [ 904.423276] env[61987]: value = "task-1061917" [ 904.423276] env[61987]: _type = "Task" [ 904.423276] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.434234] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061917, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.650121] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3b26d9f7-dac3-46a9-8582-1bddaa8c5782 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ad40937f-d5a9-4708-8b81-06087d38d765" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.252s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.722587] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061915, 'name': Rename_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.755813] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523cac0a-529b-55d5-3f88-c0aea079570f, 'name': SearchDatastore_Task, 'duration_secs': 0.008739} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.755813] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.755813] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] d11104e7-4a5c-44c3-bd92-d36c587da794/d11104e7-4a5c-44c3-bd92-d36c587da794.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 904.755813] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-247f1fd2-72bd-4ef0-ab11-8f4fb473742e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.761602] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Waiting for the task: (returnval){ [ 904.761602] env[61987]: value = "task-1061918" [ 904.761602] env[61987]: _type = "Task" [ 904.761602] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.769367] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Task: {'id': task-1061918, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.884089] env[61987]: INFO nova.compute.manager [None req-40cec40b-9e09-4a21-b7de-85cda886d5b0 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Detaching volume 41cab2ad-6048-4d47-a5db-77ff38f4f0f0 [ 904.919136] env[61987]: INFO nova.compute.manager [-] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Took 1.22 seconds to deallocate network for instance. [ 904.926431] env[61987]: INFO nova.virt.block_device [None req-40cec40b-9e09-4a21-b7de-85cda886d5b0 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Attempting to driver detach volume 41cab2ad-6048-4d47-a5db-77ff38f4f0f0 from mountpoint /dev/sdb [ 904.926431] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-40cec40b-9e09-4a21-b7de-85cda886d5b0 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Volume detach. Driver type: vmdk {{(pid=61987) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 904.926573] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-40cec40b-9e09-4a21-b7de-85cda886d5b0 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234335', 'volume_id': '41cab2ad-6048-4d47-a5db-77ff38f4f0f0', 'name': 'volume-41cab2ad-6048-4d47-a5db-77ff38f4f0f0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f937a5ec-a1d7-41d8-b998-fa18f545b304', 'attached_at': '', 'detached_at': '', 'volume_id': '41cab2ad-6048-4d47-a5db-77ff38f4f0f0', 'serial': '41cab2ad-6048-4d47-a5db-77ff38f4f0f0'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 904.931042] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b796806c-ec2c-4270-b5d6-dd59952fe7ff {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.960985] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061917, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.961704] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b6f824-8b08-4f45-8757-8a8325f6881e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.969115] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0f7b400-a873-414d-b4bb-e22f701caec8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.996345] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-183c847c-7e90-456e-8b06-1407e631c21f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.000276] env[61987]: DEBUG nova.compute.manager [req-3ac84219-7a97-4a00-a144-0a0c8e272b1c req-4a7f1206-7268-4520-a81d-46aeb4d54adb service nova] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Received event network-vif-deleted-507c557c-1a2c-4bcb-a8da-93fc1fa91bd9 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 905.017668] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-40cec40b-9e09-4a21-b7de-85cda886d5b0 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] The volume has not been displaced from its original location: [datastore2] volume-41cab2ad-6048-4d47-a5db-77ff38f4f0f0/volume-41cab2ad-6048-4d47-a5db-77ff38f4f0f0.vmdk. No consolidation needed. {{(pid=61987) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 905.023673] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-40cec40b-9e09-4a21-b7de-85cda886d5b0 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Reconfiguring VM instance instance-00000036 to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 905.024673] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb572563-bce7-45c1-8596-827592913db3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.044929] env[61987]: DEBUG oslo_vmware.api [None req-40cec40b-9e09-4a21-b7de-85cda886d5b0 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 905.044929] env[61987]: value = "task-1061919" [ 905.044929] env[61987]: _type = "Task" [ 905.044929] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.055428] env[61987]: DEBUG oslo_vmware.api [None req-40cec40b-9e09-4a21-b7de-85cda886d5b0 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061919, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.228618] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061915, 'name': Rename_Task, 'duration_secs': 1.03371} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.228929] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 905.229293] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c9571425-ee97-49d9-b0e6-577372bbfd63 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.238521] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 905.238521] env[61987]: value = "task-1061920" [ 905.238521] env[61987]: _type = "Task" [ 905.238521] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.246632] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061920, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.273904] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Task: {'id': task-1061918, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484617} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.274375] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] d11104e7-4a5c-44c3-bd92-d36c587da794/d11104e7-4a5c-44c3-bd92-d36c587da794.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 905.274780] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 905.275142] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5fbc9f3a-3177-48eb-80af-3f3143f7d483 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.281653] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Waiting for the task: (returnval){ [ 905.281653] env[61987]: value = "task-1061921" [ 905.281653] env[61987]: _type = "Task" [ 905.281653] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.295977] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Task: {'id': task-1061921, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.400910] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5401b8ce-d4ed-42b2-ab3a-8e96e83a8f71 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.408999] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f52130ab-ce0a-416e-805f-bf9f5e9d7d15 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.438911] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.442857] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cf6574d-25c8-4b23-8107-b5f203e95005 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.450994] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061917, 'name': ReconfigVM_Task, 'duration_secs': 0.524008} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.453135] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Reconfigured VM instance instance-00000052 to attach disk [datastore2] d8b9e172-c3bf-47d9-969b-7f33475d2d60/d8b9e172-c3bf-47d9-969b-7f33475d2d60.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 905.453797] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a83b8553-1ce1-46f2-913a-91dc62c5d88d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.456314] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-463630a2-66fa-43b6-bcf0-7df85aee58ea {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.471718] env[61987]: DEBUG nova.compute.provider_tree [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.474115] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for the task: (returnval){ [ 905.474115] env[61987]: value = "task-1061922" [ 905.474115] env[61987]: _type = "Task" [ 905.474115] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.484660] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061922, 'name': Rename_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.554897] env[61987]: DEBUG oslo_vmware.api [None req-40cec40b-9e09-4a21-b7de-85cda886d5b0 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061919, 'name': ReconfigVM_Task, 'duration_secs': 0.299545} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.555139] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-40cec40b-9e09-4a21-b7de-85cda886d5b0 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Reconfigured VM instance instance-00000036 to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 905.560017] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-65610428-b10a-430d-8fca-6e25de41fb32 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.575764] env[61987]: DEBUG oslo_vmware.api [None req-40cec40b-9e09-4a21-b7de-85cda886d5b0 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 905.575764] env[61987]: value = "task-1061923" [ 905.575764] env[61987]: _type = "Task" [ 905.575764] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.583823] env[61987]: DEBUG oslo_vmware.api [None req-40cec40b-9e09-4a21-b7de-85cda886d5b0 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061923, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.748961] env[61987]: DEBUG oslo_vmware.api [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061920, 'name': PowerOnVM_Task, 'duration_secs': 0.47171} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.749261] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 905.791029] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Task: {'id': task-1061921, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060284} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.791299] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 905.792056] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-950ade43-63ce-4ad8-a44a-40bc1073d6ab {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.813769] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] d11104e7-4a5c-44c3-bd92-d36c587da794/d11104e7-4a5c-44c3-bd92-d36c587da794.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 905.814057] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95717ed1-07f9-474a-96cd-d30f0eba0e1e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.833956] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Waiting for the task: (returnval){ [ 905.833956] env[61987]: value = "task-1061924" [ 905.833956] env[61987]: _type = "Task" [ 905.833956] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.841617] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Task: {'id': task-1061924, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.845637] env[61987]: DEBUG nova.compute.manager [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 905.846433] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f951da-c572-45bd-b0f4-e5a0c71022f7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.977186] env[61987]: DEBUG nova.scheduler.client.report [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 905.988285] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061922, 'name': Rename_Task, 'duration_secs': 0.135218} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.988568] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 905.988817] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0add3b29-1cf8-4e62-86d9-d21939c8e497 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.995686] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for the task: (returnval){ [ 905.995686] env[61987]: value = "task-1061925" [ 905.995686] env[61987]: _type = "Task" [ 905.995686] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.003227] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061925, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.086645] env[61987]: DEBUG oslo_vmware.api [None req-40cec40b-9e09-4a21-b7de-85cda886d5b0 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061923, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.349238] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.349238] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.361023] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Task: {'id': task-1061924, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.363350] env[61987]: DEBUG oslo_concurrency.lockutils [None req-26b797a5-0b60-4594-8eb8-ff07379a3885 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "51a36102-795e-47b7-b96a-857e54dc703e" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 34.333s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.484495] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.357s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.484626] env[61987]: DEBUG nova.compute.manager [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 906.487935] env[61987]: DEBUG oslo_concurrency.lockutils [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.251s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.488213] env[61987]: DEBUG nova.objects.instance [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lazy-loading 'resources' on Instance uuid 0422be0a-213b-41e9-b850-c823b52a90ca {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.506277] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061925, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.587069] env[61987]: DEBUG oslo_vmware.api [None req-40cec40b-9e09-4a21-b7de-85cda886d5b0 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061923, 'name': ReconfigVM_Task, 'duration_secs': 0.818459} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.587446] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-40cec40b-9e09-4a21-b7de-85cda886d5b0 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234335', 'volume_id': '41cab2ad-6048-4d47-a5db-77ff38f4f0f0', 'name': 'volume-41cab2ad-6048-4d47-a5db-77ff38f4f0f0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f937a5ec-a1d7-41d8-b998-fa18f545b304', 'attached_at': '', 'detached_at': '', 'volume_id': '41cab2ad-6048-4d47-a5db-77ff38f4f0f0', 'serial': '41cab2ad-6048-4d47-a5db-77ff38f4f0f0'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 906.849097] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Task: {'id': task-1061924, 'name': ReconfigVM_Task, 'duration_secs': 0.809752} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.850029] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Reconfigured VM instance instance-00000053 to attach disk [datastore2] d11104e7-4a5c-44c3-bd92-d36c587da794/d11104e7-4a5c-44c3-bd92-d36c587da794.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 906.850163] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4e886f2f-9e2b-472a-a82c-b93272d5dc58 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.851964] env[61987]: DEBUG nova.compute.manager [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 906.858985] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Waiting for the task: (returnval){ [ 906.858985] env[61987]: value = "task-1061926" [ 906.858985] env[61987]: _type = "Task" [ 906.858985] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.867627] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Task: {'id': task-1061926, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.989605] env[61987]: DEBUG nova.compute.utils [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 906.990989] env[61987]: DEBUG nova.compute.manager [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 906.991173] env[61987]: DEBUG nova.network.neutron [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 907.006381] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061925, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.076297] env[61987]: DEBUG nova.policy [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '697f9b1e3b3840369549a4f1ae3173ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd2afdc9bddba4ccab2b8fdb1f31da2b6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 907.140032] env[61987]: DEBUG nova.objects.instance [None req-40cec40b-9e09-4a21-b7de-85cda886d5b0 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lazy-loading 'flavor' on Instance uuid f937a5ec-a1d7-41d8-b998-fa18f545b304 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.210738] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b45bb979-cee8-4bf7-a970-9f0dcf9b4f0f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.218777] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a16a9c7e-ea08-41e6-8353-08f930ad00b9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.259982] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1521db6c-bcca-464e-8705-e53d2b9cef1e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.267699] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0597ee34-d4cc-4ffb-9717-f169a3c7014e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.280919] env[61987]: DEBUG nova.compute.provider_tree [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.374617] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Task: {'id': task-1061926, 'name': Rename_Task, 'duration_secs': 0.189885} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.375625] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.375927] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 907.376188] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-91ee0e0a-9a3c-4448-8959-86e5029a7a0c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.382416] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Waiting for the task: (returnval){ [ 907.382416] env[61987]: value = "task-1061927" [ 907.382416] env[61987]: _type = "Task" [ 907.382416] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.386706] env[61987]: DEBUG nova.network.neutron [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Successfully created port: c4833bbc-333e-46c2-a140-5e8d493d3973 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 907.393689] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Task: {'id': task-1061927, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.497281] env[61987]: DEBUG nova.compute.manager [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 907.508929] env[61987]: DEBUG oslo_vmware.api [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061925, 'name': PowerOnVM_Task, 'duration_secs': 1.036441} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.509262] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 907.509480] env[61987]: INFO nova.compute.manager [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Took 11.75 seconds to spawn the instance on the hypervisor. [ 907.509663] env[61987]: DEBUG nova.compute.manager [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 907.510463] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1417bdd5-a51f-44ec-9f56-037de22b1174 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.783795] env[61987]: DEBUG nova.scheduler.client.report [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 907.829216] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b071007-8e1e-4d03-9e36-c3afba089375 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.836128] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d8d26f96-092c-414d-bde5-767ca40c251c tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Suspending the VM {{(pid=61987) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 907.836376] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-4e92d44c-74fd-45f1-b4de-9d3ab88b2566 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.842069] env[61987]: DEBUG oslo_vmware.api [None req-d8d26f96-092c-414d-bde5-767ca40c251c tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 907.842069] env[61987]: value = "task-1061928" [ 907.842069] env[61987]: _type = "Task" [ 907.842069] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.850053] env[61987]: DEBUG oslo_vmware.api [None req-d8d26f96-092c-414d-bde5-767ca40c251c tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061928, 'name': SuspendVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.892677] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Task: {'id': task-1061927, 'name': PowerOnVM_Task} progress is 96%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.027801] env[61987]: INFO nova.compute.manager [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Took 32.55 seconds to build instance. [ 908.149398] env[61987]: DEBUG oslo_concurrency.lockutils [None req-40cec40b-9e09-4a21-b7de-85cda886d5b0 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "f937a5ec-a1d7-41d8-b998-fa18f545b304" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.768s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.289808] env[61987]: DEBUG oslo_concurrency.lockutils [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.801s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.292786] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.438s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.294753] env[61987]: INFO nova.compute.claims [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 908.317554] env[61987]: INFO nova.scheduler.client.report [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Deleted allocations for instance 0422be0a-213b-41e9-b850-c823b52a90ca [ 908.355827] env[61987]: DEBUG nova.compute.manager [req-df47f964-03ab-46f4-a56d-c6d4abbd5ffd req-44c99854-46eb-451e-858d-968545e85555 service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Received event network-changed-e81b029c-eef1-4ba6-8b98-a5647a44afe0 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 908.356035] env[61987]: DEBUG nova.compute.manager [req-df47f964-03ab-46f4-a56d-c6d4abbd5ffd req-44c99854-46eb-451e-858d-968545e85555 service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Refreshing instance network info cache due to event network-changed-e81b029c-eef1-4ba6-8b98-a5647a44afe0. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 908.356251] env[61987]: DEBUG oslo_concurrency.lockutils [req-df47f964-03ab-46f4-a56d-c6d4abbd5ffd req-44c99854-46eb-451e-858d-968545e85555 service nova] Acquiring lock "refresh_cache-d8b9e172-c3bf-47d9-969b-7f33475d2d60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.356574] env[61987]: DEBUG oslo_concurrency.lockutils [req-df47f964-03ab-46f4-a56d-c6d4abbd5ffd req-44c99854-46eb-451e-858d-968545e85555 service nova] Acquired lock "refresh_cache-d8b9e172-c3bf-47d9-969b-7f33475d2d60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.356913] env[61987]: DEBUG nova.network.neutron [req-df47f964-03ab-46f4-a56d-c6d4abbd5ffd req-44c99854-46eb-451e-858d-968545e85555 service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Refreshing network info cache for port e81b029c-eef1-4ba6-8b98-a5647a44afe0 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 908.362690] env[61987]: DEBUG oslo_vmware.api [None req-d8d26f96-092c-414d-bde5-767ca40c251c tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061928, 'name': SuspendVM_Task} progress is 58%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.392879] env[61987]: DEBUG oslo_vmware.api [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Task: {'id': task-1061927, 'name': PowerOnVM_Task, 'duration_secs': 0.602653} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.392879] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 908.393214] env[61987]: INFO nova.compute.manager [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Took 8.45 seconds to spawn the instance on the hypervisor. [ 908.393345] env[61987]: DEBUG nova.compute.manager [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 908.394046] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-934b12de-e07a-4425-bd25-64a285730c5f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.510878] env[61987]: DEBUG nova.compute.manager [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 908.533294] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5caef6ad-6a8c-4371-bff9-386cde525d7c tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "d8b9e172-c3bf-47d9-969b-7f33475d2d60" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.062s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.558137] env[61987]: DEBUG nova.virt.hardware [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 908.558374] env[61987]: DEBUG nova.virt.hardware [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 908.558597] env[61987]: DEBUG nova.virt.hardware [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 908.558858] env[61987]: DEBUG nova.virt.hardware [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 908.559028] env[61987]: DEBUG nova.virt.hardware [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 908.559193] env[61987]: DEBUG nova.virt.hardware [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 908.559409] env[61987]: DEBUG nova.virt.hardware [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 908.559579] env[61987]: DEBUG nova.virt.hardware [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 908.559832] env[61987]: DEBUG nova.virt.hardware [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 908.559926] env[61987]: DEBUG nova.virt.hardware [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 908.560155] env[61987]: DEBUG nova.virt.hardware [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 908.561506] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d078a3f9-0f12-43a3-b154-5e28f1fd980b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.569699] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7480ec04-d02a-4512-8ad1-462ae33dc669 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.683232] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "f937a5ec-a1d7-41d8-b998-fa18f545b304" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.683534] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "f937a5ec-a1d7-41d8-b998-fa18f545b304" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.683762] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "f937a5ec-a1d7-41d8-b998-fa18f545b304-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.683953] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "f937a5ec-a1d7-41d8-b998-fa18f545b304-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.684150] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "f937a5ec-a1d7-41d8-b998-fa18f545b304-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.686463] env[61987]: INFO nova.compute.manager [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Terminating instance [ 908.829114] env[61987]: DEBUG oslo_concurrency.lockutils [None req-de3f1b4e-8b8e-4a2e-883a-029b9ce0fa0a tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "0422be0a-213b-41e9-b850-c823b52a90ca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.501s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.854802] env[61987]: DEBUG oslo_vmware.api [None req-d8d26f96-092c-414d-bde5-767ca40c251c tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061928, 'name': SuspendVM_Task, 'duration_secs': 0.72088} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.854802] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d8d26f96-092c-414d-bde5-767ca40c251c tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Suspended the VM {{(pid=61987) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 908.855372] env[61987]: DEBUG nova.compute.manager [None req-d8d26f96-092c-414d-bde5-767ca40c251c tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 908.858127] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85b71016-d27f-416a-8be7-b1d8a3a0f18b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.917672] env[61987]: INFO nova.compute.manager [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Took 21.45 seconds to build instance. [ 908.994073] env[61987]: DEBUG oslo_concurrency.lockutils [None req-28fd8e91-fb83-4a46-a752-91105212a57f tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquiring lock "d8b9e172-c3bf-47d9-969b-7f33475d2d60" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.994373] env[61987]: DEBUG oslo_concurrency.lockutils [None req-28fd8e91-fb83-4a46-a752-91105212a57f tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "d8b9e172-c3bf-47d9-969b-7f33475d2d60" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.994963] env[61987]: INFO nova.compute.manager [None req-28fd8e91-fb83-4a46-a752-91105212a57f tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Rebooting instance [ 909.059170] env[61987]: DEBUG nova.compute.manager [req-8a03bfb6-7280-4513-8451-5e0b42031d7a req-369f2d0b-6c77-4284-9183-12b20d0b2de1 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Received event network-vif-plugged-c4833bbc-333e-46c2-a140-5e8d493d3973 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 909.059399] env[61987]: DEBUG oslo_concurrency.lockutils [req-8a03bfb6-7280-4513-8451-5e0b42031d7a req-369f2d0b-6c77-4284-9183-12b20d0b2de1 service nova] Acquiring lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.059619] env[61987]: DEBUG oslo_concurrency.lockutils [req-8a03bfb6-7280-4513-8451-5e0b42031d7a req-369f2d0b-6c77-4284-9183-12b20d0b2de1 service nova] Lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.059794] env[61987]: DEBUG oslo_concurrency.lockutils [req-8a03bfb6-7280-4513-8451-5e0b42031d7a req-369f2d0b-6c77-4284-9183-12b20d0b2de1 service nova] Lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.059967] env[61987]: DEBUG nova.compute.manager [req-8a03bfb6-7280-4513-8451-5e0b42031d7a req-369f2d0b-6c77-4284-9183-12b20d0b2de1 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] No waiting events found dispatching network-vif-plugged-c4833bbc-333e-46c2-a140-5e8d493d3973 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 909.060153] env[61987]: WARNING nova.compute.manager [req-8a03bfb6-7280-4513-8451-5e0b42031d7a req-369f2d0b-6c77-4284-9183-12b20d0b2de1 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Received unexpected event network-vif-plugged-c4833bbc-333e-46c2-a140-5e8d493d3973 for instance with vm_state building and task_state spawning. [ 909.144372] env[61987]: DEBUG nova.network.neutron [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Successfully updated port: c4833bbc-333e-46c2-a140-5e8d493d3973 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 909.192695] env[61987]: DEBUG nova.compute.manager [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 909.192815] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 909.193871] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e61999c-bffc-4a89-89ec-4d6d6e906986 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.201653] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 909.201889] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f5bbd9f4-ccd3-4bda-a96c-4134bb7d86e9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.207507] env[61987]: DEBUG oslo_vmware.api [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 909.207507] env[61987]: value = "task-1061929" [ 909.207507] env[61987]: _type = "Task" [ 909.207507] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.215936] env[61987]: DEBUG oslo_vmware.api [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061929, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.268863] env[61987]: DEBUG nova.network.neutron [req-df47f964-03ab-46f4-a56d-c6d4abbd5ffd req-44c99854-46eb-451e-858d-968545e85555 service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Updated VIF entry in instance network info cache for port e81b029c-eef1-4ba6-8b98-a5647a44afe0. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 909.269318] env[61987]: DEBUG nova.network.neutron [req-df47f964-03ab-46f4-a56d-c6d4abbd5ffd req-44c99854-46eb-451e-858d-968545e85555 service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Updating instance_info_cache with network_info: [{"id": "e81b029c-eef1-4ba6-8b98-a5647a44afe0", "address": "fa:16:3e:b0:72:9f", "network": {"id": "56353fe7-1959-4e8d-943c-a1ee1ffc860f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-992817851-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e43b1d992c07429685f7c89f1a8cfc9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape81b029c-ee", "ovs_interfaceid": "e81b029c-eef1-4ba6-8b98-a5647a44afe0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.418538] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a22e8140-64ad-40aa-947a-3743f3351a90 tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Lock "d11104e7-4a5c-44c3-bd92-d36c587da794" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.965s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.522906] env[61987]: DEBUG oslo_concurrency.lockutils [None req-28fd8e91-fb83-4a46-a752-91105212a57f tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquiring lock "refresh_cache-d8b9e172-c3bf-47d9-969b-7f33475d2d60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.525920] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e04fa04b-f50c-45e7-b2e2-43e21515a998 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.538021] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-773a2ef5-e276-4b87-94a7-56a237288257 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.565116] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af931a25-005f-4432-b708-4f507c9855f5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.572953] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b9fb47-06f8-425a-bb68-87bf27d9066e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.588192] env[61987]: DEBUG nova.compute.provider_tree [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.650425] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.650665] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.650925] env[61987]: DEBUG nova.network.neutron [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 909.717816] env[61987]: DEBUG oslo_vmware.api [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061929, 'name': PowerOffVM_Task, 'duration_secs': 0.199537} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.718157] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 909.718301] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 909.718562] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f67b4cbb-8304-406b-8d21-fa3eddfa984e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.778973] env[61987]: DEBUG oslo_concurrency.lockutils [req-df47f964-03ab-46f4-a56d-c6d4abbd5ffd req-44c99854-46eb-451e-858d-968545e85555 service nova] Releasing lock "refresh_cache-d8b9e172-c3bf-47d9-969b-7f33475d2d60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.779511] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 909.779750] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 909.779972] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Deleting the datastore file [datastore1] f937a5ec-a1d7-41d8-b998-fa18f545b304 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 909.780312] env[61987]: DEBUG oslo_concurrency.lockutils [None req-28fd8e91-fb83-4a46-a752-91105212a57f tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquired lock "refresh_cache-d8b9e172-c3bf-47d9-969b-7f33475d2d60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.780522] env[61987]: DEBUG nova.network.neutron [None req-28fd8e91-fb83-4a46-a752-91105212a57f tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 909.785885] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ec3d1c87-a7fe-454a-997f-34796b58fe0a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.793026] env[61987]: DEBUG oslo_vmware.api [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 909.793026] env[61987]: value = "task-1061931" [ 909.793026] env[61987]: _type = "Task" [ 909.793026] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.804644] env[61987]: DEBUG oslo_vmware.api [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061931, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.903736] env[61987]: DEBUG oslo_concurrency.lockutils [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "a720083b-7b16-468a-a595-bfe8592e8621" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.904043] env[61987]: DEBUG oslo_concurrency.lockutils [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "a720083b-7b16-468a-a595-bfe8592e8621" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.091685] env[61987]: DEBUG nova.scheduler.client.report [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 910.182976] env[61987]: DEBUG nova.network.neutron [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 910.302935] env[61987]: DEBUG oslo_vmware.api [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1061931, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133302} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.303218] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 910.303414] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 910.303600] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 910.303789] env[61987]: INFO nova.compute.manager [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Took 1.11 seconds to destroy the instance on the hypervisor. [ 910.304057] env[61987]: DEBUG oslo.service.loopingcall [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.304287] env[61987]: DEBUG nova.compute.manager [-] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 910.304371] env[61987]: DEBUG nova.network.neutron [-] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 910.331885] env[61987]: DEBUG nova.network.neutron [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Updating instance_info_cache with network_info: [{"id": "c4833bbc-333e-46c2-a140-5e8d493d3973", "address": "fa:16:3e:11:e5:35", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4833bbc-33", "ovs_interfaceid": "c4833bbc-333e-46c2-a140-5e8d493d3973", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.367244] env[61987]: INFO nova.compute.manager [None req-ca2714ca-14d0-485f-aa12-5f7088dead40 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Resuming [ 910.367867] env[61987]: DEBUG nova.objects.instance [None req-ca2714ca-14d0-485f-aa12-5f7088dead40 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lazy-loading 'flavor' on Instance uuid 51a36102-795e-47b7-b96a-857e54dc703e {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 910.381604] env[61987]: DEBUG nova.compute.manager [req-d2521893-02aa-4da4-8e74-036f00ded335 req-ef961f38-c095-4c51-9f3a-1c057267991d service nova] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Received event network-changed-0e240bad-236d-411b-b538-f0025aa3c0ea {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 910.381961] env[61987]: DEBUG nova.compute.manager [req-d2521893-02aa-4da4-8e74-036f00ded335 req-ef961f38-c095-4c51-9f3a-1c057267991d service nova] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Refreshing instance network info cache due to event network-changed-0e240bad-236d-411b-b538-f0025aa3c0ea. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 910.382047] env[61987]: DEBUG oslo_concurrency.lockutils [req-d2521893-02aa-4da4-8e74-036f00ded335 req-ef961f38-c095-4c51-9f3a-1c057267991d service nova] Acquiring lock "refresh_cache-d11104e7-4a5c-44c3-bd92-d36c587da794" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.382183] env[61987]: DEBUG oslo_concurrency.lockutils [req-d2521893-02aa-4da4-8e74-036f00ded335 req-ef961f38-c095-4c51-9f3a-1c057267991d service nova] Acquired lock "refresh_cache-d11104e7-4a5c-44c3-bd92-d36c587da794" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.382350] env[61987]: DEBUG nova.network.neutron [req-d2521893-02aa-4da4-8e74-036f00ded335 req-ef961f38-c095-4c51-9f3a-1c057267991d service nova] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Refreshing network info cache for port 0e240bad-236d-411b-b538-f0025aa3c0ea {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 910.406704] env[61987]: DEBUG nova.compute.manager [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 910.498457] env[61987]: DEBUG nova.network.neutron [None req-28fd8e91-fb83-4a46-a752-91105212a57f tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Updating instance_info_cache with network_info: [{"id": "e81b029c-eef1-4ba6-8b98-a5647a44afe0", "address": "fa:16:3e:b0:72:9f", "network": {"id": "56353fe7-1959-4e8d-943c-a1ee1ffc860f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-992817851-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e43b1d992c07429685f7c89f1a8cfc9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape81b029c-ee", "ovs_interfaceid": "e81b029c-eef1-4ba6-8b98-a5647a44afe0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.597278] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.305s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.597807] env[61987]: DEBUG nova.compute.manager [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 910.600706] env[61987]: DEBUG oslo_concurrency.lockutils [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.529s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.602234] env[61987]: INFO nova.compute.claims [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 910.835605] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.835939] env[61987]: DEBUG nova.compute.manager [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Instance network_info: |[{"id": "c4833bbc-333e-46c2-a140-5e8d493d3973", "address": "fa:16:3e:11:e5:35", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4833bbc-33", "ovs_interfaceid": "c4833bbc-333e-46c2-a140-5e8d493d3973", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 910.836373] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:11:e5:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1c8fdf9-970c-4ae0-b6d9-f1015196b552', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c4833bbc-333e-46c2-a140-5e8d493d3973', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 910.843956] env[61987]: DEBUG oslo.service.loopingcall [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.844173] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 910.844396] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f55d5dd6-efef-414b-870c-d4a8a783973c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.866839] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 910.866839] env[61987]: value = "task-1061932" [ 910.866839] env[61987]: _type = "Task" [ 910.866839] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.877470] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061932, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.935619] env[61987]: DEBUG oslo_concurrency.lockutils [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.001507] env[61987]: DEBUG oslo_concurrency.lockutils [None req-28fd8e91-fb83-4a46-a752-91105212a57f tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Releasing lock "refresh_cache-d8b9e172-c3bf-47d9-969b-7f33475d2d60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.109659] env[61987]: DEBUG nova.compute.utils [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 911.115351] env[61987]: DEBUG nova.compute.manager [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 911.115939] env[61987]: DEBUG nova.network.neutron [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 911.161244] env[61987]: DEBUG nova.compute.manager [req-0c01eaf7-4603-46dc-b24a-42366085b18b req-89b70f35-cfbb-415a-8aba-a3496b9e85f2 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Received event network-changed-c4833bbc-333e-46c2-a140-5e8d493d3973 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 911.161501] env[61987]: DEBUG nova.compute.manager [req-0c01eaf7-4603-46dc-b24a-42366085b18b req-89b70f35-cfbb-415a-8aba-a3496b9e85f2 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Refreshing instance network info cache due to event network-changed-c4833bbc-333e-46c2-a140-5e8d493d3973. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 911.161771] env[61987]: DEBUG oslo_concurrency.lockutils [req-0c01eaf7-4603-46dc-b24a-42366085b18b req-89b70f35-cfbb-415a-8aba-a3496b9e85f2 service nova] Acquiring lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.162150] env[61987]: DEBUG oslo_concurrency.lockutils [req-0c01eaf7-4603-46dc-b24a-42366085b18b req-89b70f35-cfbb-415a-8aba-a3496b9e85f2 service nova] Acquired lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.162929] env[61987]: DEBUG nova.network.neutron [req-0c01eaf7-4603-46dc-b24a-42366085b18b req-89b70f35-cfbb-415a-8aba-a3496b9e85f2 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Refreshing network info cache for port c4833bbc-333e-46c2-a140-5e8d493d3973 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 911.172609] env[61987]: DEBUG nova.policy [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e16f557dfdb04406b3fc772b891f05a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '22e62741abfb4471bbc7e11ae7d407ca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 911.198898] env[61987]: DEBUG nova.network.neutron [req-d2521893-02aa-4da4-8e74-036f00ded335 req-ef961f38-c095-4c51-9f3a-1c057267991d service nova] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Updated VIF entry in instance network info cache for port 0e240bad-236d-411b-b538-f0025aa3c0ea. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 911.199599] env[61987]: DEBUG nova.network.neutron [req-d2521893-02aa-4da4-8e74-036f00ded335 req-ef961f38-c095-4c51-9f3a-1c057267991d service nova] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Updating instance_info_cache with network_info: [{"id": "0e240bad-236d-411b-b538-f0025aa3c0ea", "address": "fa:16:3e:09:bd:1d", "network": {"id": "fa3476bc-abef-450e-9111-2e280a3f479a", "bridge": "br-int", "label": "tempest-ServersTestJSON-923853252-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.224", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e55e55f3d692423298be14eec1fd2508", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "456bd8a2-0fb6-4b17-9d25-08e7995c5184", "external-id": "nsx-vlan-transportzone-65", "segmentation_id": 65, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e240bad-23", "ovs_interfaceid": "0e240bad-236d-411b-b538-f0025aa3c0ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.306677] env[61987]: DEBUG nova.network.neutron [-] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.376721] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061932, 'name': CreateVM_Task, 'duration_secs': 0.344973} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.376916] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 911.377640] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.377818] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.378162] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 911.379738] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c9df4b5-d0e3-4f43-9341-e443f98305bd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.385580] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 911.385580] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]520e2016-c581-b304-9ff0-8645a05e2ed9" [ 911.385580] env[61987]: _type = "Task" [ 911.385580] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.394350] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]520e2016-c581-b304-9ff0-8645a05e2ed9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.422130] env[61987]: DEBUG nova.network.neutron [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Successfully created port: 09512bd5-f20a-49d6-aaf6-d4fa46cb0889 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 911.509497] env[61987]: DEBUG nova.compute.manager [None req-28fd8e91-fb83-4a46-a752-91105212a57f tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 911.510454] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-952e2085-4aa9-4f08-a2a3-6e69918b0e4e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.616907] env[61987]: DEBUG nova.compute.manager [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 911.701793] env[61987]: DEBUG oslo_concurrency.lockutils [req-d2521893-02aa-4da4-8e74-036f00ded335 req-ef961f38-c095-4c51-9f3a-1c057267991d service nova] Releasing lock "refresh_cache-d11104e7-4a5c-44c3-bd92-d36c587da794" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.809100] env[61987]: INFO nova.compute.manager [-] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Took 1.50 seconds to deallocate network for instance. [ 911.839247] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d047fc5-a9ef-4d2e-833d-8235fffa2037 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.849306] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b072573-d8b9-4475-839d-c563d0cfd54d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.882035] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea4d50ae-90db-4dbf-b2a0-70a0123f402f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.885333] env[61987]: DEBUG nova.network.neutron [req-0c01eaf7-4603-46dc-b24a-42366085b18b req-89b70f35-cfbb-415a-8aba-a3496b9e85f2 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Updated VIF entry in instance network info cache for port c4833bbc-333e-46c2-a140-5e8d493d3973. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 911.885666] env[61987]: DEBUG nova.network.neutron [req-0c01eaf7-4603-46dc-b24a-42366085b18b req-89b70f35-cfbb-415a-8aba-a3496b9e85f2 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Updating instance_info_cache with network_info: [{"id": "c4833bbc-333e-46c2-a140-5e8d493d3973", "address": "fa:16:3e:11:e5:35", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4833bbc-33", "ovs_interfaceid": "c4833bbc-333e-46c2-a140-5e8d493d3973", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.887128] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ca2714ca-14d0-485f-aa12-5f7088dead40 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "refresh_cache-51a36102-795e-47b7-b96a-857e54dc703e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.887261] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ca2714ca-14d0-485f-aa12-5f7088dead40 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquired lock "refresh_cache-51a36102-795e-47b7-b96a-857e54dc703e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.887458] env[61987]: DEBUG nova.network.neutron [None req-ca2714ca-14d0-485f-aa12-5f7088dead40 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 911.896261] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efba95eb-3d6e-4042-96c1-8a24001606c5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.903405] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]520e2016-c581-b304-9ff0-8645a05e2ed9, 'name': SearchDatastore_Task, 'duration_secs': 0.009246} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.903957] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.904202] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 911.904486] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.905281] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.905281] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 911.905281] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ed0d1c39-182a-49ea-9d29-c2a7dd3b398b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.914342] env[61987]: DEBUG nova.compute.provider_tree [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 911.922136] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 911.922335] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 911.923280] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7abd7594-2cfe-4d6f-9881-0fa117c210c6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.928514] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 911.928514] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52f1e050-300a-7af1-8644-733860d3a420" [ 911.928514] env[61987]: _type = "Task" [ 911.928514] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.936352] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52f1e050-300a-7af1-8644-733860d3a420, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.317615] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.389552] env[61987]: DEBUG oslo_concurrency.lockutils [req-0c01eaf7-4603-46dc-b24a-42366085b18b req-89b70f35-cfbb-415a-8aba-a3496b9e85f2 service nova] Releasing lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.407287] env[61987]: DEBUG nova.compute.manager [req-849c4d88-d574-45c4-980d-5926f63f13a5 req-badfc026-82c5-479f-ab41-ef5baf9301a7 service nova] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Received event network-vif-deleted-ac7175cc-c93a-4a76-aa42-faf9d1c022a1 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 912.417318] env[61987]: DEBUG nova.scheduler.client.report [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 912.439494] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52f1e050-300a-7af1-8644-733860d3a420, 'name': SearchDatastore_Task, 'duration_secs': 0.007798} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.440338] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6d106fa-691d-483c-a55c-99593ab407b5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.445497] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 912.445497] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52a480c5-286c-cfef-69b1-d351a5768119" [ 912.445497] env[61987]: _type = "Task" [ 912.445497] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.452812] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52a480c5-286c-cfef-69b1-d351a5768119, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.527143] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-934fe72d-06c1-4db9-b547-a56c5d364012 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.534728] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-28fd8e91-fb83-4a46-a752-91105212a57f tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Doing hard reboot of VM {{(pid=61987) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 912.534984] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-d9e876c6-62c3-4cc7-8274-8cd8a5ee6ea2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.541045] env[61987]: DEBUG oslo_vmware.api [None req-28fd8e91-fb83-4a46-a752-91105212a57f tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for the task: (returnval){ [ 912.541045] env[61987]: value = "task-1061933" [ 912.541045] env[61987]: _type = "Task" [ 912.541045] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.548516] env[61987]: DEBUG oslo_vmware.api [None req-28fd8e91-fb83-4a46-a752-91105212a57f tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061933, 'name': ResetVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.601408] env[61987]: DEBUG nova.network.neutron [None req-ca2714ca-14d0-485f-aa12-5f7088dead40 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Updating instance_info_cache with network_info: [{"id": "4e18ece6-5812-4c5c-827d-b3ee7d399008", "address": "fa:16:3e:1a:e6:c9", "network": {"id": "abcd2857-4c5a-4546-a003-858ddb17b0e6", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-124838733-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3fd98395e57149c987f82585296751b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "170f3b82-5915-4e36-bce9-4664ebb6be5e", "external-id": "nsx-vlan-transportzone-33", "segmentation_id": 33, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4e18ece6-58", "ovs_interfaceid": "4e18ece6-5812-4c5c-827d-b3ee7d399008", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.630273] env[61987]: DEBUG nova.compute.manager [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 912.657749] env[61987]: DEBUG nova.virt.hardware [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 912.658012] env[61987]: DEBUG nova.virt.hardware [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.658188] env[61987]: DEBUG nova.virt.hardware [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 912.658382] env[61987]: DEBUG nova.virt.hardware [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.658537] env[61987]: DEBUG nova.virt.hardware [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 912.658692] env[61987]: DEBUG nova.virt.hardware [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 912.658899] env[61987]: DEBUG nova.virt.hardware [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 912.659081] env[61987]: DEBUG nova.virt.hardware [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 912.659276] env[61987]: DEBUG nova.virt.hardware [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 912.659450] env[61987]: DEBUG nova.virt.hardware [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 912.659645] env[61987]: DEBUG nova.virt.hardware [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 912.660518] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d3edf6-0abd-4534-81d0-841db5d2658f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.668198] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01fbd8d-fb37-4d84-8a5c-479767622af2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.927037] env[61987]: DEBUG oslo_concurrency.lockutils [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.324s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.927037] env[61987]: DEBUG nova.compute.manager [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 912.928803] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.520s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.930470] env[61987]: INFO nova.compute.claims [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 912.933624] env[61987]: DEBUG nova.network.neutron [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Successfully updated port: 09512bd5-f20a-49d6-aaf6-d4fa46cb0889 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 912.957420] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52a480c5-286c-cfef-69b1-d351a5768119, 'name': SearchDatastore_Task, 'duration_secs': 0.009033} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.957782] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.958089] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] efdf62cc-9a3f-4b6b-a119-29a8096848ee/efdf62cc-9a3f-4b6b-a119-29a8096848ee.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 912.958404] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8c582728-178b-42aa-ae2d-0a4d39a47534 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.964642] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 912.964642] env[61987]: value = "task-1061934" [ 912.964642] env[61987]: _type = "Task" [ 912.964642] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.973423] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061934, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.053017] env[61987]: DEBUG oslo_vmware.api [None req-28fd8e91-fb83-4a46-a752-91105212a57f tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061933, 'name': ResetVM_Task, 'duration_secs': 0.085047} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.053017] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-28fd8e91-fb83-4a46-a752-91105212a57f tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Did hard reboot of VM {{(pid=61987) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 913.053017] env[61987]: DEBUG nova.compute.manager [None req-28fd8e91-fb83-4a46-a752-91105212a57f tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 913.053017] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37eca181-dbd9-4031-801f-e289a06295f5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.105135] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ca2714ca-14d0-485f-aa12-5f7088dead40 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Releasing lock "refresh_cache-51a36102-795e-47b7-b96a-857e54dc703e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.106306] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47e4215-a800-4b09-81ba-ae0227ce2486 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.113353] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ca2714ca-14d0-485f-aa12-5f7088dead40 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Resuming the VM {{(pid=61987) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 913.113648] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6bce1f56-6ab8-4f20-acb9-9807521e5e7a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.120954] env[61987]: DEBUG oslo_vmware.api [None req-ca2714ca-14d0-485f-aa12-5f7088dead40 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 913.120954] env[61987]: value = "task-1061935" [ 913.120954] env[61987]: _type = "Task" [ 913.120954] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.129734] env[61987]: DEBUG oslo_vmware.api [None req-ca2714ca-14d0-485f-aa12-5f7088dead40 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061935, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.203191] env[61987]: DEBUG nova.compute.manager [req-d10d20ab-0dd1-455e-830e-faa943e40593 req-139cebfa-f09a-41ea-b17e-80d438ab8bd0 service nova] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Received event network-vif-plugged-09512bd5-f20a-49d6-aaf6-d4fa46cb0889 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 913.203448] env[61987]: DEBUG oslo_concurrency.lockutils [req-d10d20ab-0dd1-455e-830e-faa943e40593 req-139cebfa-f09a-41ea-b17e-80d438ab8bd0 service nova] Acquiring lock "1b08af4a-8e3e-405e-8526-1e0a37b1f0cd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.203693] env[61987]: DEBUG oslo_concurrency.lockutils [req-d10d20ab-0dd1-455e-830e-faa943e40593 req-139cebfa-f09a-41ea-b17e-80d438ab8bd0 service nova] Lock "1b08af4a-8e3e-405e-8526-1e0a37b1f0cd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.203892] env[61987]: DEBUG oslo_concurrency.lockutils [req-d10d20ab-0dd1-455e-830e-faa943e40593 req-139cebfa-f09a-41ea-b17e-80d438ab8bd0 service nova] Lock "1b08af4a-8e3e-405e-8526-1e0a37b1f0cd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.204084] env[61987]: DEBUG nova.compute.manager [req-d10d20ab-0dd1-455e-830e-faa943e40593 req-139cebfa-f09a-41ea-b17e-80d438ab8bd0 service nova] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] No waiting events found dispatching network-vif-plugged-09512bd5-f20a-49d6-aaf6-d4fa46cb0889 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 913.204266] env[61987]: WARNING nova.compute.manager [req-d10d20ab-0dd1-455e-830e-faa943e40593 req-139cebfa-f09a-41ea-b17e-80d438ab8bd0 service nova] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Received unexpected event network-vif-plugged-09512bd5-f20a-49d6-aaf6-d4fa46cb0889 for instance with vm_state building and task_state spawning. [ 913.204549] env[61987]: DEBUG nova.compute.manager [req-d10d20ab-0dd1-455e-830e-faa943e40593 req-139cebfa-f09a-41ea-b17e-80d438ab8bd0 service nova] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Received event network-changed-09512bd5-f20a-49d6-aaf6-d4fa46cb0889 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 913.204694] env[61987]: DEBUG nova.compute.manager [req-d10d20ab-0dd1-455e-830e-faa943e40593 req-139cebfa-f09a-41ea-b17e-80d438ab8bd0 service nova] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Refreshing instance network info cache due to event network-changed-09512bd5-f20a-49d6-aaf6-d4fa46cb0889. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 913.204919] env[61987]: DEBUG oslo_concurrency.lockutils [req-d10d20ab-0dd1-455e-830e-faa943e40593 req-139cebfa-f09a-41ea-b17e-80d438ab8bd0 service nova] Acquiring lock "refresh_cache-1b08af4a-8e3e-405e-8526-1e0a37b1f0cd" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.205122] env[61987]: DEBUG oslo_concurrency.lockutils [req-d10d20ab-0dd1-455e-830e-faa943e40593 req-139cebfa-f09a-41ea-b17e-80d438ab8bd0 service nova] Acquired lock "refresh_cache-1b08af4a-8e3e-405e-8526-1e0a37b1f0cd" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.205330] env[61987]: DEBUG nova.network.neutron [req-d10d20ab-0dd1-455e-830e-faa943e40593 req-139cebfa-f09a-41ea-b17e-80d438ab8bd0 service nova] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Refreshing network info cache for port 09512bd5-f20a-49d6-aaf6-d4fa46cb0889 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 913.433086] env[61987]: DEBUG nova.compute.utils [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 913.433086] env[61987]: DEBUG nova.compute.manager [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 913.433086] env[61987]: DEBUG nova.network.neutron [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 913.444651] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "refresh_cache-1b08af4a-8e3e-405e-8526-1e0a37b1f0cd" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.475354] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061934, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.442395} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.475933] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] efdf62cc-9a3f-4b6b-a119-29a8096848ee/efdf62cc-9a3f-4b6b-a119-29a8096848ee.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 913.476816] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 913.477241] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5979c62e-26ac-45ab-95d2-fa54a123e024 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.485127] env[61987]: DEBUG nova.policy [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c95828500a3446eb4a6c20a5c966b23', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '38d4c69cad9d40309abf5885ad36939e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 913.487081] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 913.487081] env[61987]: value = "task-1061936" [ 913.487081] env[61987]: _type = "Task" [ 913.487081] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.496453] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061936, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.566202] env[61987]: DEBUG oslo_concurrency.lockutils [None req-28fd8e91-fb83-4a46-a752-91105212a57f tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "d8b9e172-c3bf-47d9-969b-7f33475d2d60" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.571s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.637107] env[61987]: DEBUG oslo_vmware.api [None req-ca2714ca-14d0-485f-aa12-5f7088dead40 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061935, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.763085] env[61987]: DEBUG nova.network.neutron [req-d10d20ab-0dd1-455e-830e-faa943e40593 req-139cebfa-f09a-41ea-b17e-80d438ab8bd0 service nova] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 913.781727] env[61987]: DEBUG nova.network.neutron [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Successfully created port: a60eaa2f-cdd3-4b07-9a40-d48a4dd95d17 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 913.940751] env[61987]: DEBUG nova.compute.manager [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 913.998827] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061936, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.241341} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.001563] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 914.002896] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98cda5b3-a25d-41f3-954f-60427e80612f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.030089] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] efdf62cc-9a3f-4b6b-a119-29a8096848ee/efdf62cc-9a3f-4b6b-a119-29a8096848ee.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 914.034511] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac3aa3f7-c09a-4233-bd5a-ebfd81e635e4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.055634] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 914.055634] env[61987]: value = "task-1061937" [ 914.055634] env[61987]: _type = "Task" [ 914.055634] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.068525] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061937, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.087941] env[61987]: DEBUG nova.network.neutron [req-d10d20ab-0dd1-455e-830e-faa943e40593 req-139cebfa-f09a-41ea-b17e-80d438ab8bd0 service nova] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.131830] env[61987]: DEBUG oslo_vmware.api [None req-ca2714ca-14d0-485f-aa12-5f7088dead40 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061935, 'name': PowerOnVM_Task, 'duration_secs': 0.59541} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.134201] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ca2714ca-14d0-485f-aa12-5f7088dead40 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Resumed the VM {{(pid=61987) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 914.134415] env[61987]: DEBUG nova.compute.manager [None req-ca2714ca-14d0-485f-aa12-5f7088dead40 tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 914.135473] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30ae1f33-797a-4a53-9f94-7c60eb5be33d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.214810] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69e84c03-b770-4101-841a-7b6b0ff7ca69 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.221986] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2dca87d-82ec-43cc-90bb-f7c616dd574f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.255051] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa4cff30-8a25-4ece-847e-891dc3e70a25 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.262847] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10580db9-470e-480b-a265-1376d63444da {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.277153] env[61987]: DEBUG nova.compute.provider_tree [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.568080] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061937, 'name': ReconfigVM_Task, 'duration_secs': 0.242414} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.569170] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Reconfigured VM instance instance-00000054 to attach disk [datastore1] efdf62cc-9a3f-4b6b-a119-29a8096848ee/efdf62cc-9a3f-4b6b-a119-29a8096848ee.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.573149] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bd084eae-4d46-452c-9172-29f52ad3cc18 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.583139] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 914.583139] env[61987]: value = "task-1061938" [ 914.583139] env[61987]: _type = "Task" [ 914.583139] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.590760] env[61987]: DEBUG oslo_concurrency.lockutils [req-d10d20ab-0dd1-455e-830e-faa943e40593 req-139cebfa-f09a-41ea-b17e-80d438ab8bd0 service nova] Releasing lock "refresh_cache-1b08af4a-8e3e-405e-8526-1e0a37b1f0cd" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.591189] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061938, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.591384] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired lock "refresh_cache-1b08af4a-8e3e-405e-8526-1e0a37b1f0cd" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.591547] env[61987]: DEBUG nova.network.neutron [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 914.644687] env[61987]: DEBUG nova.compute.manager [req-c02000bd-d55d-4c58-a7c2-4b77ebc17b3e req-c69870f2-cbaa-4b56-a43a-b179f8e3d58b service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Received event network-changed-e81b029c-eef1-4ba6-8b98-a5647a44afe0 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 914.644970] env[61987]: DEBUG nova.compute.manager [req-c02000bd-d55d-4c58-a7c2-4b77ebc17b3e req-c69870f2-cbaa-4b56-a43a-b179f8e3d58b service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Refreshing instance network info cache due to event network-changed-e81b029c-eef1-4ba6-8b98-a5647a44afe0. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 914.645503] env[61987]: DEBUG oslo_concurrency.lockutils [req-c02000bd-d55d-4c58-a7c2-4b77ebc17b3e req-c69870f2-cbaa-4b56-a43a-b179f8e3d58b service nova] Acquiring lock "refresh_cache-d8b9e172-c3bf-47d9-969b-7f33475d2d60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.645872] env[61987]: DEBUG oslo_concurrency.lockutils [req-c02000bd-d55d-4c58-a7c2-4b77ebc17b3e req-c69870f2-cbaa-4b56-a43a-b179f8e3d58b service nova] Acquired lock "refresh_cache-d8b9e172-c3bf-47d9-969b-7f33475d2d60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.646028] env[61987]: DEBUG nova.network.neutron [req-c02000bd-d55d-4c58-a7c2-4b77ebc17b3e req-c69870f2-cbaa-4b56-a43a-b179f8e3d58b service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Refreshing network info cache for port e81b029c-eef1-4ba6-8b98-a5647a44afe0 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 914.781263] env[61987]: DEBUG nova.scheduler.client.report [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 914.950747] env[61987]: DEBUG nova.compute.manager [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 914.976826] env[61987]: DEBUG nova.virt.hardware [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 914.977188] env[61987]: DEBUG nova.virt.hardware [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 914.977416] env[61987]: DEBUG nova.virt.hardware [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 914.977679] env[61987]: DEBUG nova.virt.hardware [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 914.977893] env[61987]: DEBUG nova.virt.hardware [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 914.978128] env[61987]: DEBUG nova.virt.hardware [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 914.978426] env[61987]: DEBUG nova.virt.hardware [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 914.978658] env[61987]: DEBUG nova.virt.hardware [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 914.978905] env[61987]: DEBUG nova.virt.hardware [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 914.979159] env[61987]: DEBUG nova.virt.hardware [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 914.979412] env[61987]: DEBUG nova.virt.hardware [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 914.980568] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee8dd417-acc6-4494-ac5a-52bf6b0441a2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.992627] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a404a5a-4375-4e30-bb7d-ac66304d93a3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.093373] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061938, 'name': Rename_Task, 'duration_secs': 0.147209} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.095499] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 915.095777] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f4c3cacc-5e5f-4da6-8231-c7fe7e679129 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.102522] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 915.102522] env[61987]: value = "task-1061939" [ 915.102522] env[61987]: _type = "Task" [ 915.102522] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.110268] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061939, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.125158] env[61987]: DEBUG nova.network.neutron [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 915.266607] env[61987]: DEBUG nova.compute.manager [req-e1784cb8-f039-464d-b171-5600aa11c191 req-1224d282-ffbf-444a-96f5-348e731acf4f service nova] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Received event network-vif-plugged-a60eaa2f-cdd3-4b07-9a40-d48a4dd95d17 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 915.266834] env[61987]: DEBUG oslo_concurrency.lockutils [req-e1784cb8-f039-464d-b171-5600aa11c191 req-1224d282-ffbf-444a-96f5-348e731acf4f service nova] Acquiring lock "5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.267114] env[61987]: DEBUG oslo_concurrency.lockutils [req-e1784cb8-f039-464d-b171-5600aa11c191 req-1224d282-ffbf-444a-96f5-348e731acf4f service nova] Lock "5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.267250] env[61987]: DEBUG oslo_concurrency.lockutils [req-e1784cb8-f039-464d-b171-5600aa11c191 req-1224d282-ffbf-444a-96f5-348e731acf4f service nova] Lock "5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.267387] env[61987]: DEBUG nova.compute.manager [req-e1784cb8-f039-464d-b171-5600aa11c191 req-1224d282-ffbf-444a-96f5-348e731acf4f service nova] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] No waiting events found dispatching network-vif-plugged-a60eaa2f-cdd3-4b07-9a40-d48a4dd95d17 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 915.267551] env[61987]: WARNING nova.compute.manager [req-e1784cb8-f039-464d-b171-5600aa11c191 req-1224d282-ffbf-444a-96f5-348e731acf4f service nova] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Received unexpected event network-vif-plugged-a60eaa2f-cdd3-4b07-9a40-d48a4dd95d17 for instance with vm_state building and task_state spawning. [ 915.285928] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.357s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.286461] env[61987]: DEBUG nova.compute.manager [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 915.288946] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquiring lock "d8b9e172-c3bf-47d9-969b-7f33475d2d60" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.289570] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "d8b9e172-c3bf-47d9-969b-7f33475d2d60" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.289570] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquiring lock "d8b9e172-c3bf-47d9-969b-7f33475d2d60-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.289767] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "d8b9e172-c3bf-47d9-969b-7f33475d2d60-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.289830] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "d8b9e172-c3bf-47d9-969b-7f33475d2d60-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.295688] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.488s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.298167] env[61987]: INFO nova.compute.claims [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 915.305512] env[61987]: INFO nova.compute.manager [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Terminating instance [ 915.384167] env[61987]: DEBUG nova.network.neutron [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Successfully updated port: a60eaa2f-cdd3-4b07-9a40-d48a4dd95d17 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 915.429209] env[61987]: DEBUG nova.network.neutron [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Updating instance_info_cache with network_info: [{"id": "09512bd5-f20a-49d6-aaf6-d4fa46cb0889", "address": "fa:16:3e:8e:12:bf", "network": {"id": "d5ed1e9e-9033-441e-bfcb-2ac643d7f03f", "bridge": "br-int", "label": "tempest-ServersTestJSON-8466993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e62741abfb4471bbc7e11ae7d407ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39ab9baf-90cd-4fe2-8d56-434f8210fc19", "external-id": "nsx-vlan-transportzone-713", "segmentation_id": 713, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap09512bd5-f2", "ovs_interfaceid": "09512bd5-f20a-49d6-aaf6-d4fa46cb0889", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.535409] env[61987]: DEBUG nova.network.neutron [req-c02000bd-d55d-4c58-a7c2-4b77ebc17b3e req-c69870f2-cbaa-4b56-a43a-b179f8e3d58b service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Updated VIF entry in instance network info cache for port e81b029c-eef1-4ba6-8b98-a5647a44afe0. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 915.535826] env[61987]: DEBUG nova.network.neutron [req-c02000bd-d55d-4c58-a7c2-4b77ebc17b3e req-c69870f2-cbaa-4b56-a43a-b179f8e3d58b service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Updating instance_info_cache with network_info: [{"id": "e81b029c-eef1-4ba6-8b98-a5647a44afe0", "address": "fa:16:3e:b0:72:9f", "network": {"id": "56353fe7-1959-4e8d-943c-a1ee1ffc860f", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-992817851-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e43b1d992c07429685f7c89f1a8cfc9a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "93c5b7ce-4c84-40bc-884c-b2453e0eee69", "external-id": "nsx-vlan-transportzone-882", "segmentation_id": 882, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape81b029c-ee", "ovs_interfaceid": "e81b029c-eef1-4ba6-8b98-a5647a44afe0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.613627] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061939, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.797496] env[61987]: DEBUG nova.compute.utils [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 915.799351] env[61987]: DEBUG nova.compute.manager [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 915.799616] env[61987]: DEBUG nova.network.neutron [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 915.818881] env[61987]: DEBUG nova.compute.manager [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 915.819219] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 915.820921] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f20c410-e42d-483b-861f-c8f840e83dc8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.829077] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 915.829315] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80635d8a-7a1f-48a3-ad09-9c5fc212cdb6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.835746] env[61987]: DEBUG oslo_vmware.api [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for the task: (returnval){ [ 915.835746] env[61987]: value = "task-1061940" [ 915.835746] env[61987]: _type = "Task" [ 915.835746] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.844253] env[61987]: DEBUG oslo_vmware.api [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061940, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.845810] env[61987]: DEBUG nova.policy [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c95828500a3446eb4a6c20a5c966b23', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '38d4c69cad9d40309abf5885ad36939e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 915.889975] env[61987]: DEBUG oslo_concurrency.lockutils [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "refresh_cache-5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.890172] env[61987]: DEBUG oslo_concurrency.lockutils [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquired lock "refresh_cache-5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.890603] env[61987]: DEBUG nova.network.neutron [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 915.936265] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Releasing lock "refresh_cache-1b08af4a-8e3e-405e-8526-1e0a37b1f0cd" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.936682] env[61987]: DEBUG nova.compute.manager [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Instance network_info: |[{"id": "09512bd5-f20a-49d6-aaf6-d4fa46cb0889", "address": "fa:16:3e:8e:12:bf", "network": {"id": "d5ed1e9e-9033-441e-bfcb-2ac643d7f03f", "bridge": "br-int", "label": "tempest-ServersTestJSON-8466993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e62741abfb4471bbc7e11ae7d407ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39ab9baf-90cd-4fe2-8d56-434f8210fc19", "external-id": "nsx-vlan-transportzone-713", "segmentation_id": 713, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap09512bd5-f2", "ovs_interfaceid": "09512bd5-f20a-49d6-aaf6-d4fa46cb0889", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 915.937912] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:12:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '39ab9baf-90cd-4fe2-8d56-434f8210fc19', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '09512bd5-f20a-49d6-aaf6-d4fa46cb0889', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 915.948542] env[61987]: DEBUG oslo.service.loopingcall [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.948791] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 915.951560] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-32eb37b5-f78b-439d-80ea-ab2025438e78 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.969495] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 915.969495] env[61987]: value = "task-1061941" [ 915.969495] env[61987]: _type = "Task" [ 915.969495] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.977809] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061941, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.038518] env[61987]: DEBUG oslo_concurrency.lockutils [req-c02000bd-d55d-4c58-a7c2-4b77ebc17b3e req-c69870f2-cbaa-4b56-a43a-b179f8e3d58b service nova] Releasing lock "refresh_cache-d8b9e172-c3bf-47d9-969b-7f33475d2d60" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.115856] env[61987]: DEBUG oslo_vmware.api [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1061939, 'name': PowerOnVM_Task, 'duration_secs': 0.660945} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.116288] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 916.116498] env[61987]: INFO nova.compute.manager [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Took 7.61 seconds to spawn the instance on the hypervisor. [ 916.116701] env[61987]: DEBUG nova.compute.manager [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 916.117519] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30dd1f60-b629-42a9-b11d-048f946b0cf6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.199759] env[61987]: DEBUG nova.network.neutron [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Successfully created port: dfa5785c-1dca-49b8-88ed-a067c0dfb815 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 916.303300] env[61987]: DEBUG nova.compute.manager [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 916.346478] env[61987]: DEBUG oslo_vmware.api [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061940, 'name': PowerOffVM_Task, 'duration_secs': 0.306787} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.349213] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 916.349397] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 916.350059] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fd00a646-dd74-4048-98b4-e621bd0ceac1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.421876] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 916.422184] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 916.422476] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Deleting the datastore file [datastore2] d8b9e172-c3bf-47d9-969b-7f33475d2d60 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 916.423680] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ee29ca44-7b14-4645-a0d5-28cfc2cc3785 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.428500] env[61987]: DEBUG oslo_vmware.api [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for the task: (returnval){ [ 916.428500] env[61987]: value = "task-1061943" [ 916.428500] env[61987]: _type = "Task" [ 916.428500] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.438193] env[61987]: DEBUG oslo_vmware.api [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061943, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.452636] env[61987]: DEBUG nova.network.neutron [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 916.480848] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061941, 'name': CreateVM_Task, 'duration_secs': 0.450102} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.483365] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 916.484120] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.488019] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.488019] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 916.488019] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1a3f9b4-2779-4ad7-b2bc-7f9225cbba86 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.492729] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 916.492729] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52b0cc91-2994-f552-921a-82ebccd3231c" [ 916.492729] env[61987]: _type = "Task" [ 916.492729] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.502026] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b0cc91-2994-f552-921a-82ebccd3231c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.615230] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fab7715-6940-42e0-bc9f-71727c1f7213 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.627018] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-756bc86a-59cb-4eb1-8913-b5f42df61fcf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.663787] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51457091-930d-4d6d-baf6-e96c348b5a4d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.666787] env[61987]: INFO nova.compute.manager [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Took 25.50 seconds to build instance. [ 916.675419] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfa07690-fb16-4b50-94a1-eb5c968c9bd8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.692028] env[61987]: DEBUG nova.compute.provider_tree [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.742246] env[61987]: DEBUG nova.network.neutron [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Updating instance_info_cache with network_info: [{"id": "a60eaa2f-cdd3-4b07-9a40-d48a4dd95d17", "address": "fa:16:3e:8f:7f:02", "network": {"id": "c167f42e-b158-462b-966f-d744475140cc", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1189472144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38d4c69cad9d40309abf5885ad36939e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "22b8c642-38ad-4c11-9051-145ab3bc54f2", "external-id": "nsx-vlan-transportzone-247", "segmentation_id": 247, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa60eaa2f-cd", "ovs_interfaceid": "a60eaa2f-cdd3-4b07-9a40-d48a4dd95d17", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.938233] env[61987]: DEBUG oslo_vmware.api [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061943, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169633} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.938523] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 916.938705] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 916.938883] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 916.939074] env[61987]: INFO nova.compute.manager [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Took 1.12 seconds to destroy the instance on the hypervisor. [ 916.939329] env[61987]: DEBUG oslo.service.loopingcall [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 916.939523] env[61987]: DEBUG nova.compute.manager [-] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 916.939620] env[61987]: DEBUG nova.network.neutron [-] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 917.003668] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b0cc91-2994-f552-921a-82ebccd3231c, 'name': SearchDatastore_Task, 'duration_secs': 0.010159} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.003957] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.004212] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 917.004459] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.004641] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.004834] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 917.005115] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-674b71b2-2463-4c29-999d-1ef1004d492c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.013384] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 917.013575] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 917.014329] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfa0aa0b-3749-4497-ab18-a00d6c09d502 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.019626] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 917.019626] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52cf4931-1fd3-eb9f-a847-4e4d329856fd" [ 917.019626] env[61987]: _type = "Task" [ 917.019626] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.028187] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52cf4931-1fd3-eb9f-a847-4e4d329856fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.171087] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2c32e42-0677-4ce8-ba4d-66e6d9d23893 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.011s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.195403] env[61987]: DEBUG nova.scheduler.client.report [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 917.245099] env[61987]: DEBUG oslo_concurrency.lockutils [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Releasing lock "refresh_cache-5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.245543] env[61987]: DEBUG nova.compute.manager [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Instance network_info: |[{"id": "a60eaa2f-cdd3-4b07-9a40-d48a4dd95d17", "address": "fa:16:3e:8f:7f:02", "network": {"id": "c167f42e-b158-462b-966f-d744475140cc", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1189472144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38d4c69cad9d40309abf5885ad36939e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "22b8c642-38ad-4c11-9051-145ab3bc54f2", "external-id": "nsx-vlan-transportzone-247", "segmentation_id": 247, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa60eaa2f-cd", "ovs_interfaceid": "a60eaa2f-cdd3-4b07-9a40-d48a4dd95d17", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 917.246334] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8f:7f:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '22b8c642-38ad-4c11-9051-145ab3bc54f2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a60eaa2f-cdd3-4b07-9a40-d48a4dd95d17', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 917.254199] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Creating folder: Project (38d4c69cad9d40309abf5885ad36939e). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 917.254521] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-14311c4f-886b-4445-9246-63370da8b2b5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.268465] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Created folder: Project (38d4c69cad9d40309abf5885ad36939e) in parent group-v234219. [ 917.268465] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Creating folder: Instances. Parent ref: group-v234382. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 917.268465] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5ce3ede2-b6aa-4456-93bd-c01abbf43fe7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.276599] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Created folder: Instances in parent group-v234382. [ 917.276841] env[61987]: DEBUG oslo.service.loopingcall [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 917.277044] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 917.277891] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f1b79bb1-b10b-432d-9915-d0740a0c11ff {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.296244] env[61987]: DEBUG nova.compute.manager [req-5d8e7e4a-7461-48ee-bf83-efe44aa1763e req-51559bea-4e5b-401e-b077-498eb59ceffc service nova] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Received event network-changed-a60eaa2f-cdd3-4b07-9a40-d48a4dd95d17 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 917.296244] env[61987]: DEBUG nova.compute.manager [req-5d8e7e4a-7461-48ee-bf83-efe44aa1763e req-51559bea-4e5b-401e-b077-498eb59ceffc service nova] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Refreshing instance network info cache due to event network-changed-a60eaa2f-cdd3-4b07-9a40-d48a4dd95d17. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 917.296244] env[61987]: DEBUG oslo_concurrency.lockutils [req-5d8e7e4a-7461-48ee-bf83-efe44aa1763e req-51559bea-4e5b-401e-b077-498eb59ceffc service nova] Acquiring lock "refresh_cache-5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.296244] env[61987]: DEBUG oslo_concurrency.lockutils [req-5d8e7e4a-7461-48ee-bf83-efe44aa1763e req-51559bea-4e5b-401e-b077-498eb59ceffc service nova] Acquired lock "refresh_cache-5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.296244] env[61987]: DEBUG nova.network.neutron [req-5d8e7e4a-7461-48ee-bf83-efe44aa1763e req-51559bea-4e5b-401e-b077-498eb59ceffc service nova] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Refreshing network info cache for port a60eaa2f-cdd3-4b07-9a40-d48a4dd95d17 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 917.299916] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 917.299916] env[61987]: value = "task-1061946" [ 917.299916] env[61987]: _type = "Task" [ 917.299916] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.311670] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061946, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.313182] env[61987]: DEBUG nova.compute.manager [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 917.343473] env[61987]: DEBUG nova.virt.hardware [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 917.344597] env[61987]: DEBUG nova.virt.hardware [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 917.344597] env[61987]: DEBUG nova.virt.hardware [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 917.344597] env[61987]: DEBUG nova.virt.hardware [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 917.344597] env[61987]: DEBUG nova.virt.hardware [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 917.344818] env[61987]: DEBUG nova.virt.hardware [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 917.345373] env[61987]: DEBUG nova.virt.hardware [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 917.345474] env[61987]: DEBUG nova.virt.hardware [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 917.345687] env[61987]: DEBUG nova.virt.hardware [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 917.345893] env[61987]: DEBUG nova.virt.hardware [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 917.346123] env[61987]: DEBUG nova.virt.hardware [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 917.347469] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43294015-b0fc-4f59-9924-2e1f023a6f1f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.356709] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-923bbb1d-4621-498f-bb71-131834c7fd84 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.482905] env[61987]: DEBUG nova.compute.manager [req-007c35b6-fe00-4575-9f70-d0e5610db2b0 req-3369d580-95ba-45bb-96ea-05c38982d4b5 service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Received event network-vif-deleted-e81b029c-eef1-4ba6-8b98-a5647a44afe0 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 917.483138] env[61987]: INFO nova.compute.manager [req-007c35b6-fe00-4575-9f70-d0e5610db2b0 req-3369d580-95ba-45bb-96ea-05c38982d4b5 service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Neutron deleted interface e81b029c-eef1-4ba6-8b98-a5647a44afe0; detaching it from the instance and deleting it from the info cache [ 917.483320] env[61987]: DEBUG nova.network.neutron [req-007c35b6-fe00-4575-9f70-d0e5610db2b0 req-3369d580-95ba-45bb-96ea-05c38982d4b5 service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.530158] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52cf4931-1fd3-eb9f-a847-4e4d329856fd, 'name': SearchDatastore_Task, 'duration_secs': 0.014352} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.530906] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf73a379-b7c8-4c79-a2a3-94aa6a010fb4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.536782] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 917.536782] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]523c388b-2b9f-19e4-e136-618e1f0b74cf" [ 917.536782] env[61987]: _type = "Task" [ 917.536782] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.544701] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523c388b-2b9f-19e4-e136-618e1f0b74cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.701081] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.405s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.701081] env[61987]: DEBUG nova.compute.manager [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 917.705568] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.477s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.705827] env[61987]: DEBUG nova.objects.instance [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lazy-loading 'resources' on Instance uuid 212e4860-751e-43f7-80c6-25b6784e3541 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.814223] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061946, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.948192] env[61987]: DEBUG nova.network.neutron [-] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.988069] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bb093934-be0c-445b-ac40-ccf9044a5c92 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.998140] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-504ff9f4-12b1-4268-917c-efc54574823c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.029332] env[61987]: DEBUG nova.compute.manager [req-007c35b6-fe00-4575-9f70-d0e5610db2b0 req-3369d580-95ba-45bb-96ea-05c38982d4b5 service nova] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Detach interface failed, port_id=e81b029c-eef1-4ba6-8b98-a5647a44afe0, reason: Instance d8b9e172-c3bf-47d9-969b-7f33475d2d60 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 918.050023] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523c388b-2b9f-19e4-e136-618e1f0b74cf, 'name': SearchDatastore_Task, 'duration_secs': 0.013273} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.050023] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.050023] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd/1b08af4a-8e3e-405e-8526-1e0a37b1f0cd.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 918.050023] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-af8e3400-3f88-47ce-a750-c92b63c891ee {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.054718] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 918.054718] env[61987]: value = "task-1061947" [ 918.054718] env[61987]: _type = "Task" [ 918.054718] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.062571] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061947, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.128263] env[61987]: DEBUG nova.network.neutron [req-5d8e7e4a-7461-48ee-bf83-efe44aa1763e req-51559bea-4e5b-401e-b077-498eb59ceffc service nova] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Updated VIF entry in instance network info cache for port a60eaa2f-cdd3-4b07-9a40-d48a4dd95d17. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 918.128638] env[61987]: DEBUG nova.network.neutron [req-5d8e7e4a-7461-48ee-bf83-efe44aa1763e req-51559bea-4e5b-401e-b077-498eb59ceffc service nova] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Updating instance_info_cache with network_info: [{"id": "a60eaa2f-cdd3-4b07-9a40-d48a4dd95d17", "address": "fa:16:3e:8f:7f:02", "network": {"id": "c167f42e-b158-462b-966f-d744475140cc", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1189472144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38d4c69cad9d40309abf5885ad36939e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "22b8c642-38ad-4c11-9051-145ab3bc54f2", "external-id": "nsx-vlan-transportzone-247", "segmentation_id": 247, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa60eaa2f-cd", "ovs_interfaceid": "a60eaa2f-cdd3-4b07-9a40-d48a4dd95d17", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.133754] env[61987]: DEBUG nova.network.neutron [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Successfully updated port: dfa5785c-1dca-49b8-88ed-a067c0dfb815 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 918.209246] env[61987]: DEBUG nova.compute.utils [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 918.214658] env[61987]: DEBUG nova.compute.manager [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 918.217093] env[61987]: DEBUG nova.network.neutron [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 918.254419] env[61987]: DEBUG nova.policy [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c95828500a3446eb4a6c20a5c966b23', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '38d4c69cad9d40309abf5885ad36939e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 918.311714] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061946, 'name': CreateVM_Task, 'duration_secs': 0.53192} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.314316] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 918.315154] env[61987]: DEBUG oslo_concurrency.lockutils [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.315333] env[61987]: DEBUG oslo_concurrency.lockutils [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.315804] env[61987]: DEBUG oslo_concurrency.lockutils [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 918.315932] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-645c2571-f153-41aa-b8ac-6bc831a2e5dc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.320668] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 918.320668] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52d64a49-ff61-6d55-9e07-cce686d968df" [ 918.320668] env[61987]: _type = "Task" [ 918.320668] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.330750] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52d64a49-ff61-6d55-9e07-cce686d968df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.452206] env[61987]: INFO nova.compute.manager [-] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Took 1.51 seconds to deallocate network for instance. [ 918.480048] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a11dd8b-f638-4295-90bf-9770a2b30752 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.489460] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba30c0c-0af9-4d20-9495-fd07e4d9c4df {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.527694] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-993de4a1-f1e0-491f-88d6-b3b7343f7963 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.536649] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2431c715-a318-4f5b-9c65-a53a336a9fc3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.552916] env[61987]: DEBUG nova.compute.provider_tree [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.554612] env[61987]: DEBUG nova.network.neutron [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Successfully created port: acb22031-7e46-4139-8097-e06110e2fb54 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 918.567912] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061947, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.632713] env[61987]: DEBUG oslo_concurrency.lockutils [req-5d8e7e4a-7461-48ee-bf83-efe44aa1763e req-51559bea-4e5b-401e-b077-498eb59ceffc service nova] Releasing lock "refresh_cache-5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.634950] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "refresh_cache-286f24dd-2239-434e-8ae7-15164a0ea8b0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.635140] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquired lock "refresh_cache-286f24dd-2239-434e-8ae7-15164a0ea8b0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.635301] env[61987]: DEBUG nova.network.neutron [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 918.718739] env[61987]: DEBUG nova.compute.manager [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 918.837819] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52d64a49-ff61-6d55-9e07-cce686d968df, 'name': SearchDatastore_Task, 'duration_secs': 0.012837} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.838256] env[61987]: DEBUG oslo_concurrency.lockutils [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.838498] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 918.838741] env[61987]: DEBUG oslo_concurrency.lockutils [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.838923] env[61987]: DEBUG oslo_concurrency.lockutils [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.839131] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 918.839580] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-895838aa-a880-4353-acdd-7e6bf93890e5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.848808] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 918.849065] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 918.849985] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69465464-8fa4-44a0-8858-0e6ceb761dd3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.856466] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 918.856466] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]524dfea4-29c9-29d5-faba-a972f508450c" [ 918.856466] env[61987]: _type = "Task" [ 918.856466] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.864710] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524dfea4-29c9-29d5-faba-a972f508450c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.960833] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.057921] env[61987]: DEBUG nova.scheduler.client.report [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 919.070433] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061947, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.742376} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.070834] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd/1b08af4a-8e3e-405e-8526-1e0a37b1f0cd.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 919.071168] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 919.071484] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f42910c5-e9c4-4a0a-bb21-24615ee780fb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.078501] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 919.078501] env[61987]: value = "task-1061948" [ 919.078501] env[61987]: _type = "Task" [ 919.078501] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.086433] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061948, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.169033] env[61987]: DEBUG nova.network.neutron [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 919.339584] env[61987]: DEBUG nova.network.neutron [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Updating instance_info_cache with network_info: [{"id": "dfa5785c-1dca-49b8-88ed-a067c0dfb815", "address": "fa:16:3e:d8:24:06", "network": {"id": "c167f42e-b158-462b-966f-d744475140cc", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1189472144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38d4c69cad9d40309abf5885ad36939e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "22b8c642-38ad-4c11-9051-145ab3bc54f2", "external-id": "nsx-vlan-transportzone-247", "segmentation_id": 247, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdfa5785c-1d", "ovs_interfaceid": "dfa5785c-1dca-49b8-88ed-a067c0dfb815", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.369031] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524dfea4-29c9-29d5-faba-a972f508450c, 'name': SearchDatastore_Task, 'duration_secs': 0.011601} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.369814] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d97d25a8-1038-47f9-965f-b551e24f2478 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.375771] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 919.375771] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5270eba7-c4a1-5947-82d4-64972b47c656" [ 919.375771] env[61987]: _type = "Task" [ 919.375771] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.383320] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5270eba7-c4a1-5947-82d4-64972b47c656, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.510473] env[61987]: DEBUG nova.compute.manager [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Received event network-changed-c4833bbc-333e-46c2-a140-5e8d493d3973 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 919.510706] env[61987]: DEBUG nova.compute.manager [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Refreshing instance network info cache due to event network-changed-c4833bbc-333e-46c2-a140-5e8d493d3973. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 919.510937] env[61987]: DEBUG oslo_concurrency.lockutils [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] Acquiring lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.511113] env[61987]: DEBUG oslo_concurrency.lockutils [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] Acquired lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.511296] env[61987]: DEBUG nova.network.neutron [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Refreshing network info cache for port c4833bbc-333e-46c2-a140-5e8d493d3973 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 919.566252] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.860s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.568498] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 17.279s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.588987] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061948, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.159886} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.589962] env[61987]: INFO nova.scheduler.client.report [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Deleted allocations for instance 212e4860-751e-43f7-80c6-25b6784e3541 [ 919.591015] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 919.593854] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f0c6c4e-9d11-46fc-87b1-18e219f3deb9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.616680] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd/1b08af4a-8e3e-405e-8526-1e0a37b1f0cd.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 919.618114] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-246dc299-f867-4860-875c-12add61ccfb8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.640713] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 919.640713] env[61987]: value = "task-1061949" [ 919.640713] env[61987]: _type = "Task" [ 919.640713] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.649130] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061949, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.730397] env[61987]: DEBUG nova.compute.manager [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 919.756506] env[61987]: DEBUG nova.virt.hardware [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 919.756828] env[61987]: DEBUG nova.virt.hardware [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 919.757028] env[61987]: DEBUG nova.virt.hardware [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 919.757284] env[61987]: DEBUG nova.virt.hardware [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 919.757481] env[61987]: DEBUG nova.virt.hardware [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 919.757640] env[61987]: DEBUG nova.virt.hardware [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 919.757853] env[61987]: DEBUG nova.virt.hardware [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 919.758033] env[61987]: DEBUG nova.virt.hardware [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 919.758214] env[61987]: DEBUG nova.virt.hardware [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 919.758409] env[61987]: DEBUG nova.virt.hardware [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 919.758632] env[61987]: DEBUG nova.virt.hardware [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 919.759530] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63533220-ec14-4225-825e-4491ad95414b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.768374] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bf317a4-c7fc-4172-9fd1-f1be2704b3d5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.842288] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Releasing lock "refresh_cache-286f24dd-2239-434e-8ae7-15164a0ea8b0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.842699] env[61987]: DEBUG nova.compute.manager [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Instance network_info: |[{"id": "dfa5785c-1dca-49b8-88ed-a067c0dfb815", "address": "fa:16:3e:d8:24:06", "network": {"id": "c167f42e-b158-462b-966f-d744475140cc", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1189472144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38d4c69cad9d40309abf5885ad36939e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "22b8c642-38ad-4c11-9051-145ab3bc54f2", "external-id": "nsx-vlan-transportzone-247", "segmentation_id": 247, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdfa5785c-1d", "ovs_interfaceid": "dfa5785c-1dca-49b8-88ed-a067c0dfb815", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 919.843236] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:24:06', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '22b8c642-38ad-4c11-9051-145ab3bc54f2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dfa5785c-1dca-49b8-88ed-a067c0dfb815', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 919.852114] env[61987]: DEBUG oslo.service.loopingcall [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 919.852514] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 919.852598] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-756385ea-fb13-44da-ab26-8b8ee4c29788 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.875387] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 919.875387] env[61987]: value = "task-1061950" [ 919.875387] env[61987]: _type = "Task" [ 919.875387] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.887957] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5270eba7-c4a1-5947-82d4-64972b47c656, 'name': SearchDatastore_Task, 'duration_secs': 0.072965} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.891641] env[61987]: DEBUG oslo_concurrency.lockutils [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.891966] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32/5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 919.892184] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061950, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.892406] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a5d89767-52e6-4407-aa25-02576d7165b5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.898856] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 919.898856] env[61987]: value = "task-1061951" [ 919.898856] env[61987]: _type = "Task" [ 919.898856] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.910220] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061951, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.062574] env[61987]: DEBUG oslo_concurrency.lockutils [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "51a36102-795e-47b7-b96a-857e54dc703e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.062926] env[61987]: DEBUG oslo_concurrency.lockutils [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "51a36102-795e-47b7-b96a-857e54dc703e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.063171] env[61987]: DEBUG oslo_concurrency.lockutils [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "51a36102-795e-47b7-b96a-857e54dc703e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.063513] env[61987]: DEBUG oslo_concurrency.lockutils [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "51a36102-795e-47b7-b96a-857e54dc703e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.063712] env[61987]: DEBUG oslo_concurrency.lockutils [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "51a36102-795e-47b7-b96a-857e54dc703e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.065880] env[61987]: INFO nova.compute.manager [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Terminating instance [ 920.099954] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6fd5325-4383-443a-b247-3f6062edcd94 tempest-ImagesTestJSON-932483618 tempest-ImagesTestJSON-932483618-project-member] Lock "212e4860-751e-43f7-80c6-25b6784e3541" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.404s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.152012] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061949, 'name': ReconfigVM_Task, 'duration_secs': 0.485873} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.152317] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd/1b08af4a-8e3e-405e-8526-1e0a37b1f0cd.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 920.153063] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-43139c1f-1336-49ac-b40b-5d4e103eff35 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.159815] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 920.159815] env[61987]: value = "task-1061952" [ 920.159815] env[61987]: _type = "Task" [ 920.159815] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.171064] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061952, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.389921] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061950, 'name': CreateVM_Task, 'duration_secs': 0.340639} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.390319] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 920.391458] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.391821] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.392450] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 920.392882] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-022b6d0e-c1e7-46e9-9595-2552fbdab20b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.400105] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 920.400105] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5262266b-2f95-44f2-3429-5699c85b171a" [ 920.400105] env[61987]: _type = "Task" [ 920.400105] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.420288] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061951, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.425985] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5262266b-2f95-44f2-3429-5699c85b171a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.570291] env[61987]: DEBUG nova.compute.manager [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 920.570586] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 920.571581] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-110a2238-7bd6-4054-bc4e-4c9d109fa28d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.583793] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 920.584061] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a88fdb1-6e96-42fe-86ad-e2dd3862dab9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.593365] env[61987]: DEBUG oslo_vmware.api [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 920.593365] env[61987]: value = "task-1061953" [ 920.593365] env[61987]: _type = "Task" [ 920.593365] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.604123] env[61987]: DEBUG oslo_vmware.api [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061953, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.609690] env[61987]: WARNING nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance f937a5ec-a1d7-41d8-b998-fa18f545b304 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 920.609881] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 84a4a69b-04db-4546-ba89-9039e382a0c4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.610073] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.610243] env[61987]: WARNING nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 920.610314] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 69a22e40-d469-4500-926e-0a12a233f252 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.610435] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 51a36102-795e-47b7-b96a-857e54dc703e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.610592] env[61987]: WARNING nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance d8b9e172-c3bf-47d9-969b-7f33475d2d60 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 920.610713] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance d11104e7-4a5c-44c3-bd92-d36c587da794 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.610837] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance efdf62cc-9a3f-4b6b-a119-29a8096848ee actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.610952] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.611107] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.611241] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 286f24dd-2239-434e-8ae7-15164a0ea8b0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.611346] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 876b74ae-cae8-44a9-96ba-e2869ed43be6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 920.670584] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061952, 'name': Rename_Task, 'duration_secs': 0.462939} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.670913] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 920.671443] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-97f4ea4d-73b2-4129-bef4-08551db346cf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.677593] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 920.677593] env[61987]: value = "task-1061954" [ 920.677593] env[61987]: _type = "Task" [ 920.677593] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.685707] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061954, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.698330] env[61987]: DEBUG nova.network.neutron [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Successfully updated port: acb22031-7e46-4139-8097-e06110e2fb54 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 920.727491] env[61987]: DEBUG nova.network.neutron [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Updated VIF entry in instance network info cache for port c4833bbc-333e-46c2-a140-5e8d493d3973. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 920.728104] env[61987]: DEBUG nova.network.neutron [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Updating instance_info_cache with network_info: [{"id": "c4833bbc-333e-46c2-a140-5e8d493d3973", "address": "fa:16:3e:11:e5:35", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4833bbc-33", "ovs_interfaceid": "c4833bbc-333e-46c2-a140-5e8d493d3973", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.912465] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061951, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.582785} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.912899] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32/5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 920.913145] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 920.913405] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fdbc1abe-6ed7-42b4-a5b1-66e95366253c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.919205] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5262266b-2f95-44f2-3429-5699c85b171a, 'name': SearchDatastore_Task, 'duration_secs': 0.025061} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.919816] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.920106] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 920.920315] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.920512] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.920708] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 920.920954] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-51a3dc5f-7eb1-49d4-88f0-b797f43e942b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.923661] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 920.923661] env[61987]: value = "task-1061955" [ 920.923661] env[61987]: _type = "Task" [ 920.923661] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.932090] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061955, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.933214] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 920.933399] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 920.934146] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5fa9abd2-485e-4fd1-b32a-924380f202e5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.939558] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 920.939558] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]521107fa-e9b3-74d1-b5e6-c85c3710cd29" [ 920.939558] env[61987]: _type = "Task" [ 920.939558] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.947101] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]521107fa-e9b3-74d1-b5e6-c85c3710cd29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.104368] env[61987]: DEBUG oslo_vmware.api [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061953, 'name': PowerOffVM_Task, 'duration_secs': 0.210218} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.104953] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 921.105370] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 921.106060] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cdf5727e-68f3-4db5-b36d-89719b9e3607 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.114666] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance ad9041d9-1e0f-46d6-93f7-efe82197deb6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 921.188518] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061954, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.200626] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "refresh_cache-876b74ae-cae8-44a9-96ba-e2869ed43be6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.200802] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquired lock "refresh_cache-876b74ae-cae8-44a9-96ba-e2869ed43be6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.200971] env[61987]: DEBUG nova.network.neutron [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 921.230801] env[61987]: DEBUG oslo_concurrency.lockutils [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] Releasing lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.231035] env[61987]: DEBUG nova.compute.manager [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Received event network-vif-plugged-dfa5785c-1dca-49b8-88ed-a067c0dfb815 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 921.231284] env[61987]: DEBUG oslo_concurrency.lockutils [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] Acquiring lock "286f24dd-2239-434e-8ae7-15164a0ea8b0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.231526] env[61987]: DEBUG oslo_concurrency.lockutils [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] Lock "286f24dd-2239-434e-8ae7-15164a0ea8b0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.231729] env[61987]: DEBUG oslo_concurrency.lockutils [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] Lock "286f24dd-2239-434e-8ae7-15164a0ea8b0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.231920] env[61987]: DEBUG nova.compute.manager [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] No waiting events found dispatching network-vif-plugged-dfa5785c-1dca-49b8-88ed-a067c0dfb815 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 921.232137] env[61987]: WARNING nova.compute.manager [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Received unexpected event network-vif-plugged-dfa5785c-1dca-49b8-88ed-a067c0dfb815 for instance with vm_state building and task_state spawning. [ 921.232344] env[61987]: DEBUG nova.compute.manager [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Received event network-changed-dfa5785c-1dca-49b8-88ed-a067c0dfb815 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 921.232661] env[61987]: DEBUG nova.compute.manager [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Refreshing instance network info cache due to event network-changed-dfa5785c-1dca-49b8-88ed-a067c0dfb815. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 921.232772] env[61987]: DEBUG oslo_concurrency.lockutils [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] Acquiring lock "refresh_cache-286f24dd-2239-434e-8ae7-15164a0ea8b0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.232940] env[61987]: DEBUG oslo_concurrency.lockutils [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] Acquired lock "refresh_cache-286f24dd-2239-434e-8ae7-15164a0ea8b0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.233137] env[61987]: DEBUG nova.network.neutron [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Refreshing network info cache for port dfa5785c-1dca-49b8-88ed-a067c0dfb815 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 921.437708] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061955, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066859} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.437708] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 921.437708] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f02027-2e37-4a84-ae8c-2d46dcb7080f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.463014] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32/5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 921.463868] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-438d6fa5-7e82-4787-9d3d-5ff16bccd121 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.483032] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]521107fa-e9b3-74d1-b5e6-c85c3710cd29, 'name': SearchDatastore_Task, 'duration_secs': 0.020213} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.484381] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c57ee9a7-9e9f-46bc-909f-0c4339d251ba {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.488890] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 921.488890] env[61987]: value = "task-1061957" [ 921.488890] env[61987]: _type = "Task" [ 921.488890] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.490078] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 921.490078] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]526df363-5ec1-4fdd-d980-fff5d51c9bf6" [ 921.490078] env[61987]: _type = "Task" [ 921.490078] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.501278] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]526df363-5ec1-4fdd-d980-fff5d51c9bf6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.504341] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061957, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.534134] env[61987]: DEBUG nova.compute.manager [req-ab2b9f3a-1b36-468e-af33-ed00f6fe1017 req-e33e5946-a99f-4bb7-ac81-6dbbdf7c74a9 service nova] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Received event network-vif-plugged-acb22031-7e46-4139-8097-e06110e2fb54 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 921.534366] env[61987]: DEBUG oslo_concurrency.lockutils [req-ab2b9f3a-1b36-468e-af33-ed00f6fe1017 req-e33e5946-a99f-4bb7-ac81-6dbbdf7c74a9 service nova] Acquiring lock "876b74ae-cae8-44a9-96ba-e2869ed43be6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.534636] env[61987]: DEBUG oslo_concurrency.lockutils [req-ab2b9f3a-1b36-468e-af33-ed00f6fe1017 req-e33e5946-a99f-4bb7-ac81-6dbbdf7c74a9 service nova] Lock "876b74ae-cae8-44a9-96ba-e2869ed43be6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.534819] env[61987]: DEBUG oslo_concurrency.lockutils [req-ab2b9f3a-1b36-468e-af33-ed00f6fe1017 req-e33e5946-a99f-4bb7-ac81-6dbbdf7c74a9 service nova] Lock "876b74ae-cae8-44a9-96ba-e2869ed43be6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.535032] env[61987]: DEBUG nova.compute.manager [req-ab2b9f3a-1b36-468e-af33-ed00f6fe1017 req-e33e5946-a99f-4bb7-ac81-6dbbdf7c74a9 service nova] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] No waiting events found dispatching network-vif-plugged-acb22031-7e46-4139-8097-e06110e2fb54 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 921.535782] env[61987]: WARNING nova.compute.manager [req-ab2b9f3a-1b36-468e-af33-ed00f6fe1017 req-e33e5946-a99f-4bb7-ac81-6dbbdf7c74a9 service nova] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Received unexpected event network-vif-plugged-acb22031-7e46-4139-8097-e06110e2fb54 for instance with vm_state building and task_state spawning. [ 921.535975] env[61987]: DEBUG nova.compute.manager [req-ab2b9f3a-1b36-468e-af33-ed00f6fe1017 req-e33e5946-a99f-4bb7-ac81-6dbbdf7c74a9 service nova] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Received event network-changed-acb22031-7e46-4139-8097-e06110e2fb54 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 921.536163] env[61987]: DEBUG nova.compute.manager [req-ab2b9f3a-1b36-468e-af33-ed00f6fe1017 req-e33e5946-a99f-4bb7-ac81-6dbbdf7c74a9 service nova] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Refreshing instance network info cache due to event network-changed-acb22031-7e46-4139-8097-e06110e2fb54. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 921.536339] env[61987]: DEBUG oslo_concurrency.lockutils [req-ab2b9f3a-1b36-468e-af33-ed00f6fe1017 req-e33e5946-a99f-4bb7-ac81-6dbbdf7c74a9 service nova] Acquiring lock "refresh_cache-876b74ae-cae8-44a9-96ba-e2869ed43be6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.594977] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 921.595774] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 921.595851] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Deleting the datastore file [datastore2] 51a36102-795e-47b7-b96a-857e54dc703e {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.596222] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b42d1efe-ef24-4774-bd34-4c9fd18ebf8a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.607773] env[61987]: DEBUG oslo_vmware.api [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for the task: (returnval){ [ 921.607773] env[61987]: value = "task-1061958" [ 921.607773] env[61987]: _type = "Task" [ 921.607773] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.616051] env[61987]: DEBUG oslo_vmware.api [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061958, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.617838] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance ae1396b3-c28b-4d1c-9a3a-3dc2170847ca has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 921.690362] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061954, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.750249] env[61987]: DEBUG nova.network.neutron [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 922.004777] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061957, 'name': ReconfigVM_Task, 'duration_secs': 0.439617} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.009376] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32/5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 922.010556] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]526df363-5ec1-4fdd-d980-fff5d51c9bf6, 'name': SearchDatastore_Task, 'duration_secs': 0.078743} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.010903] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-183e3530-f03f-4235-bb3b-39662d4f054d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.013289] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.013701] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 286f24dd-2239-434e-8ae7-15164a0ea8b0/286f24dd-2239-434e-8ae7-15164a0ea8b0.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 922.014243] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5b3fb9b4-59f6-4ce9-817a-036c5a611add {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.027675] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 922.027675] env[61987]: value = "task-1061960" [ 922.027675] env[61987]: _type = "Task" [ 922.027675] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.027955] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 922.027955] env[61987]: value = "task-1061959" [ 922.027955] env[61987]: _type = "Task" [ 922.027955] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.042017] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061959, 'name': Rename_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.044553] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061960, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.109400] env[61987]: DEBUG nova.network.neutron [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Updating instance_info_cache with network_info: [{"id": "acb22031-7e46-4139-8097-e06110e2fb54", "address": "fa:16:3e:3c:fc:6a", "network": {"id": "c167f42e-b158-462b-966f-d744475140cc", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1189472144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38d4c69cad9d40309abf5885ad36939e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "22b8c642-38ad-4c11-9051-145ab3bc54f2", "external-id": "nsx-vlan-transportzone-247", "segmentation_id": 247, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapacb22031-7e", "ovs_interfaceid": "acb22031-7e46-4139-8097-e06110e2fb54", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.121417] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance a720083b-7b16-468a-a595-bfe8592e8621 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 922.121691] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 922.121835] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2496MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 922.124324] env[61987]: DEBUG oslo_vmware.api [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Task: {'id': task-1061958, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.246242} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.126017] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 922.126017] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 922.126017] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 922.126017] env[61987]: INFO nova.compute.manager [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Took 1.56 seconds to destroy the instance on the hypervisor. [ 922.126530] env[61987]: DEBUG oslo.service.loopingcall [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.126530] env[61987]: DEBUG nova.compute.manager [-] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 922.126530] env[61987]: DEBUG nova.network.neutron [-] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 922.191667] env[61987]: DEBUG oslo_vmware.api [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061954, 'name': PowerOnVM_Task, 'duration_secs': 1.407881} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.192204] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 922.192432] env[61987]: INFO nova.compute.manager [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Took 9.56 seconds to spawn the instance on the hypervisor. [ 922.192638] env[61987]: DEBUG nova.compute.manager [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 922.193488] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33b56b3d-632e-4fc4-bde7-a7caf0d2dbb6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.196927] env[61987]: DEBUG nova.network.neutron [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Updated VIF entry in instance network info cache for port dfa5785c-1dca-49b8-88ed-a067c0dfb815. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 922.197305] env[61987]: DEBUG nova.network.neutron [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Updating instance_info_cache with network_info: [{"id": "dfa5785c-1dca-49b8-88ed-a067c0dfb815", "address": "fa:16:3e:d8:24:06", "network": {"id": "c167f42e-b158-462b-966f-d744475140cc", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1189472144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38d4c69cad9d40309abf5885ad36939e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "22b8c642-38ad-4c11-9051-145ab3bc54f2", "external-id": "nsx-vlan-transportzone-247", "segmentation_id": 247, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdfa5785c-1d", "ovs_interfaceid": "dfa5785c-1dca-49b8-88ed-a067c0dfb815", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.400043] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a603840-77c0-47e9-ba44-554eb1eb209f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.415047] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10e5204b-578f-4394-b33c-adf5d3ee9834 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.450354] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a7c0232-1def-4421-90f5-4b83e69b532a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.459971] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c89d0206-59cd-494f-9dba-8c1f403b83e0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.474860] env[61987]: DEBUG nova.compute.provider_tree [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.542852] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061960, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.515348} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.545909] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 286f24dd-2239-434e-8ae7-15164a0ea8b0/286f24dd-2239-434e-8ae7-15164a0ea8b0.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 922.546162] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 922.546441] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061959, 'name': Rename_Task, 'duration_secs': 0.149791} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.546661] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-931a1814-4be8-4aee-b7a8-4f4e2ea6b4f2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.548677] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 922.549191] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1f3d8b5b-383a-445f-98bc-75b4290067bf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.556608] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 922.556608] env[61987]: value = "task-1061961" [ 922.556608] env[61987]: _type = "Task" [ 922.556608] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.561023] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 922.561023] env[61987]: value = "task-1061962" [ 922.561023] env[61987]: _type = "Task" [ 922.561023] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.568655] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061961, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.573401] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061962, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.616668] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Releasing lock "refresh_cache-876b74ae-cae8-44a9-96ba-e2869ed43be6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.617123] env[61987]: DEBUG nova.compute.manager [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Instance network_info: |[{"id": "acb22031-7e46-4139-8097-e06110e2fb54", "address": "fa:16:3e:3c:fc:6a", "network": {"id": "c167f42e-b158-462b-966f-d744475140cc", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1189472144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38d4c69cad9d40309abf5885ad36939e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "22b8c642-38ad-4c11-9051-145ab3bc54f2", "external-id": "nsx-vlan-transportzone-247", "segmentation_id": 247, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapacb22031-7e", "ovs_interfaceid": "acb22031-7e46-4139-8097-e06110e2fb54", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 922.617500] env[61987]: DEBUG oslo_concurrency.lockutils [req-ab2b9f3a-1b36-468e-af33-ed00f6fe1017 req-e33e5946-a99f-4bb7-ac81-6dbbdf7c74a9 service nova] Acquired lock "refresh_cache-876b74ae-cae8-44a9-96ba-e2869ed43be6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.617835] env[61987]: DEBUG nova.network.neutron [req-ab2b9f3a-1b36-468e-af33-ed00f6fe1017 req-e33e5946-a99f-4bb7-ac81-6dbbdf7c74a9 service nova] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Refreshing network info cache for port acb22031-7e46-4139-8097-e06110e2fb54 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 922.619898] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:fc:6a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '22b8c642-38ad-4c11-9051-145ab3bc54f2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'acb22031-7e46-4139-8097-e06110e2fb54', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 922.630067] env[61987]: DEBUG oslo.service.loopingcall [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.631689] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 922.633024] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ddb03e6-568f-4a09-8c5c-a738511576a3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.653434] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 922.653434] env[61987]: value = "task-1061963" [ 922.653434] env[61987]: _type = "Task" [ 922.653434] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.661691] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061963, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.700339] env[61987]: DEBUG oslo_concurrency.lockutils [req-ac67cb16-6464-4b1c-bd79-db23a44aa014 req-9b3e74d3-6774-40f1-8172-05a1800c8e8d service nova] Releasing lock "refresh_cache-286f24dd-2239-434e-8ae7-15164a0ea8b0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.716329] env[61987]: INFO nova.compute.manager [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Took 23.88 seconds to build instance. [ 922.915189] env[61987]: DEBUG nova.network.neutron [-] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.978135] env[61987]: DEBUG nova.scheduler.client.report [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 923.069279] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061961, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075821} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.070169] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 923.070963] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-825873be-060a-408c-9c31-b4728583aa3b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.076819] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061962, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.095947] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 286f24dd-2239-434e-8ae7-15164a0ea8b0/286f24dd-2239-434e-8ae7-15164a0ea8b0.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.096256] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66b3a998-34c6-49b8-9f61-e293e78be830 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.115094] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 923.115094] env[61987]: value = "task-1061964" [ 923.115094] env[61987]: _type = "Task" [ 923.115094] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.124747] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061964, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.165186] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061963, 'name': CreateVM_Task, 'duration_secs': 0.365081} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.165408] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 923.166064] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.166248] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.166617] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 923.166886] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e49ef54e-583e-4ffe-b96d-7f58164b56a0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.171628] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 923.171628] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52b834d2-7e11-3b24-9544-8a5a5e53a694" [ 923.171628] env[61987]: _type = "Task" [ 923.171628] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.180193] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b834d2-7e11-3b24-9544-8a5a5e53a694, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.219097] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f6d83a10-424a-4892-9df0-bdd181cf1e01 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "1b08af4a-8e3e-405e-8526-1e0a37b1f0cd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.396s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.368170] env[61987]: DEBUG nova.network.neutron [req-ab2b9f3a-1b36-468e-af33-ed00f6fe1017 req-e33e5946-a99f-4bb7-ac81-6dbbdf7c74a9 service nova] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Updated VIF entry in instance network info cache for port acb22031-7e46-4139-8097-e06110e2fb54. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 923.368576] env[61987]: DEBUG nova.network.neutron [req-ab2b9f3a-1b36-468e-af33-ed00f6fe1017 req-e33e5946-a99f-4bb7-ac81-6dbbdf7c74a9 service nova] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Updating instance_info_cache with network_info: [{"id": "acb22031-7e46-4139-8097-e06110e2fb54", "address": "fa:16:3e:3c:fc:6a", "network": {"id": "c167f42e-b158-462b-966f-d744475140cc", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1189472144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38d4c69cad9d40309abf5885ad36939e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "22b8c642-38ad-4c11-9051-145ab3bc54f2", "external-id": "nsx-vlan-transportzone-247", "segmentation_id": 247, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapacb22031-7e", "ovs_interfaceid": "acb22031-7e46-4139-8097-e06110e2fb54", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.418012] env[61987]: INFO nova.compute.manager [-] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Took 1.29 seconds to deallocate network for instance. [ 923.483518] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61987) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 923.483898] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.916s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.484416] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.819s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.486027] env[61987]: INFO nova.compute.claims [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 923.572579] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061962, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.627156] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061964, 'name': ReconfigVM_Task, 'duration_secs': 0.307894} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.627517] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 286f24dd-2239-434e-8ae7-15164a0ea8b0/286f24dd-2239-434e-8ae7-15164a0ea8b0.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 923.628513] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1c76f2cc-997f-446e-87e1-e16ffd27f59f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.636866] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 923.636866] env[61987]: value = "task-1061965" [ 923.636866] env[61987]: _type = "Task" [ 923.636866] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.643903] env[61987]: DEBUG nova.compute.manager [req-792aeaa8-3e37-43cb-9a61-7b61ba2debd9 req-789afc31-cfe1-4ce6-9f25-a7bd06e4f60a service nova] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Received event network-vif-deleted-4e18ece6-5812-4c5c-827d-b3ee7d399008 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 923.649804] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061965, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.682752] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b834d2-7e11-3b24-9544-8a5a5e53a694, 'name': SearchDatastore_Task, 'duration_secs': 0.019701} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.683090] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.683335] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 923.683637] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.683824] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.684033] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 923.684459] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c1d5697e-8dc9-4675-ae11-2b6e37033002 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.694764] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 923.694971] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 923.696785] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e3df2f5-157d-435b-b553-17c7860103ee {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.702427] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 923.702427] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52872df3-d062-7327-a16f-79528d45d341" [ 923.702427] env[61987]: _type = "Task" [ 923.702427] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.712153] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52872df3-d062-7327-a16f-79528d45d341, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.871363] env[61987]: DEBUG oslo_concurrency.lockutils [req-ab2b9f3a-1b36-468e-af33-ed00f6fe1017 req-e33e5946-a99f-4bb7-ac81-6dbbdf7c74a9 service nova] Releasing lock "refresh_cache-876b74ae-cae8-44a9-96ba-e2869ed43be6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.926060] env[61987]: DEBUG oslo_concurrency.lockutils [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.073187] env[61987]: DEBUG oslo_vmware.api [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061962, 'name': PowerOnVM_Task, 'duration_secs': 1.265438} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.073686] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 924.074064] env[61987]: INFO nova.compute.manager [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Took 9.12 seconds to spawn the instance on the hypervisor. [ 924.074626] env[61987]: DEBUG nova.compute.manager [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 924.077368] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be680805-ca9e-4521-954f-f9649aa803dc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.148214] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061965, 'name': Rename_Task, 'duration_secs': 0.142978} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.148214] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 924.148214] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3a6ba515-27c0-4f77-88a7-391cf012eb04 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.155515] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 924.155515] env[61987]: value = "task-1061966" [ 924.155515] env[61987]: _type = "Task" [ 924.155515] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.166300] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061966, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.197108] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ff40f32-de33-48ac-81a3-a3001773cf2a tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "69a22e40-d469-4500-926e-0a12a233f252" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.197407] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ff40f32-de33-48ac-81a3-a3001773cf2a tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "69a22e40-d469-4500-926e-0a12a233f252" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.197688] env[61987]: DEBUG nova.compute.manager [None req-8ff40f32-de33-48ac-81a3-a3001773cf2a tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 924.198523] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b8016dd-dbe0-4222-ae02-bdb741df9c52 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.208532] env[61987]: DEBUG nova.compute.manager [None req-8ff40f32-de33-48ac-81a3-a3001773cf2a tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61987) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 924.209201] env[61987]: DEBUG nova.objects.instance [None req-8ff40f32-de33-48ac-81a3-a3001773cf2a tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lazy-loading 'flavor' on Instance uuid 69a22e40-d469-4500-926e-0a12a233f252 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.216851] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52872df3-d062-7327-a16f-79528d45d341, 'name': SearchDatastore_Task, 'duration_secs': 0.017455} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.217648] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88bf3003-51b3-4fab-8350-45b135c9e12b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.224346] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 924.224346] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52a53678-9f79-c04b-fef6-b76bae8e6fc9" [ 924.224346] env[61987]: _type = "Task" [ 924.224346] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.235117] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52a53678-9f79-c04b-fef6-b76bae8e6fc9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.595326] env[61987]: INFO nova.compute.manager [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Took 25.55 seconds to build instance. [ 924.667727] env[61987]: DEBUG oslo_vmware.api [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061966, 'name': PowerOnVM_Task, 'duration_secs': 0.458449} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.668059] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 924.668279] env[61987]: INFO nova.compute.manager [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Took 7.35 seconds to spawn the instance on the hypervisor. [ 924.668608] env[61987]: DEBUG nova.compute.manager [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 924.669531] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f238b5f-2281-4fe4-ae95-dd5623bfb497 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.738373] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52a53678-9f79-c04b-fef6-b76bae8e6fc9, 'name': SearchDatastore_Task, 'duration_secs': 0.012283} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.738695] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.738763] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 876b74ae-cae8-44a9-96ba-e2869ed43be6/876b74ae-cae8-44a9-96ba-e2869ed43be6.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 924.740046] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c4f7778b-bb97-4005-9f50-1a57dcd8dc5d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.746631] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 924.746631] env[61987]: value = "task-1061967" [ 924.746631] env[61987]: _type = "Task" [ 924.746631] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.754982] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "f9c82f38-7e06-4949-80a5-6eaa3c01c88c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.755274] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "f9c82f38-7e06-4949-80a5-6eaa3c01c88c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.756881] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca4853c7-4a0c-4572-91a9-e1d7607cc600 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.763193] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061967, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.769428] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24a3558f-c2aa-440b-af3f-4f6832db5676 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.804447] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58612bc3-59be-4f78-8724-32f38d756934 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.812254] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a93aca7-3d3e-453c-9b89-5ebab61b7d71 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.826715] env[61987]: DEBUG nova.compute.provider_tree [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.100228] env[61987]: DEBUG oslo_concurrency.lockutils [None req-162fe639-6860-4032-9761-951dd4fee681 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.067s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.193051] env[61987]: INFO nova.compute.manager [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Took 25.81 seconds to build instance. [ 925.217012] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ff40f32-de33-48ac-81a3-a3001773cf2a tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 925.217518] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fa0ef4f7-b0a9-458d-8402-1aad585c3650 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.236162] env[61987]: DEBUG oslo_vmware.api [None req-8ff40f32-de33-48ac-81a3-a3001773cf2a tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 925.236162] env[61987]: value = "task-1061968" [ 925.236162] env[61987]: _type = "Task" [ 925.236162] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.250325] env[61987]: DEBUG oslo_vmware.api [None req-8ff40f32-de33-48ac-81a3-a3001773cf2a tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1061968, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.260584] env[61987]: DEBUG nova.compute.manager [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 925.263934] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061967, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.330169] env[61987]: DEBUG nova.scheduler.client.report [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 925.699409] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1d2f9a44-ee01-4f95-8b4f-009de5336e31 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "286f24dd-2239-434e-8ae7-15164a0ea8b0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.322s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.751062] env[61987]: DEBUG oslo_vmware.api [None req-8ff40f32-de33-48ac-81a3-a3001773cf2a tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1061968, 'name': PowerOffVM_Task, 'duration_secs': 0.30082} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.751062] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ff40f32-de33-48ac-81a3-a3001773cf2a tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 925.751062] env[61987]: DEBUG nova.compute.manager [None req-8ff40f32-de33-48ac-81a3-a3001773cf2a tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 925.751062] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e86cc7f7-d67e-47b9-8941-00bfa5fc590a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.762893] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061967, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.558304} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.763414] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 876b74ae-cae8-44a9-96ba-e2869ed43be6/876b74ae-cae8-44a9-96ba-e2869ed43be6.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 925.763787] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 925.764048] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d2dfdc64-9bb4-4ccd-aea3-b1b61bf87544 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.773012] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 925.773012] env[61987]: value = "task-1061969" [ 925.773012] env[61987]: _type = "Task" [ 925.773012] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.781357] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061969, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.793308] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.835821] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.351s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.836210] env[61987]: DEBUG nova.compute.manager [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 925.839126] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.400s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.839337] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.841343] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.466s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.843263] env[61987]: INFO nova.compute.claims [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 925.872514] env[61987]: INFO nova.scheduler.client.report [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Deleted allocations for instance ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05 [ 926.269035] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8ff40f32-de33-48ac-81a3-a3001773cf2a tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "69a22e40-d469-4500-926e-0a12a233f252" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.071s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.284236] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061969, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.235493} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.285027] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 926.287130] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4588b571-ad9f-445f-91f5-1a6cb30d351f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.319345] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 876b74ae-cae8-44a9-96ba-e2869ed43be6/876b74ae-cae8-44a9-96ba-e2869ed43be6.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 926.319345] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b196b084-188f-459a-9511-c0e21680db59 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.342443] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 926.342443] env[61987]: value = "task-1061970" [ 926.342443] env[61987]: _type = "Task" [ 926.342443] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.349130] env[61987]: DEBUG nova.compute.utils [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 926.353023] env[61987]: DEBUG nova.compute.manager [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 926.353023] env[61987]: DEBUG nova.network.neutron [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 926.361809] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061970, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.383644] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2e7a44ae-75ee-49f1-96ac-5d0bf0e172b7 tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.387s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.400558] env[61987]: DEBUG nova.policy [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4ead7955f1d475590523ac98c579956', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '315a2a3450024ed38e27cd7758f392f9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 926.759289] env[61987]: DEBUG nova.objects.instance [None req-121f5830-0aaf-4714-bbd9-ccf079025ad9 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lazy-loading 'flavor' on Instance uuid 69a22e40-d469-4500-926e-0a12a233f252 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.772350] env[61987]: DEBUG nova.network.neutron [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Successfully created port: d857e05f-901f-4f86-9f9e-33aed6d5fec9 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 926.857383] env[61987]: DEBUG nova.compute.manager [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 926.863484] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061970, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.079781] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beddeebf-571a-4aea-a6e0-4f07a4854557 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.087559] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbff5221-f738-4119-9f94-fb5c32f53a91 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.119941] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ccdd87f-084c-4053-a014-0643f78c22cf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.127265] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-072295e8-a9bf-4a6e-904d-1282f728a9f2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.140641] env[61987]: DEBUG nova.compute.provider_tree [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.265541] env[61987]: DEBUG oslo_concurrency.lockutils [None req-121f5830-0aaf-4714-bbd9-ccf079025ad9 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.265808] env[61987]: DEBUG oslo_concurrency.lockutils [None req-121f5830-0aaf-4714-bbd9-ccf079025ad9 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.265914] env[61987]: DEBUG nova.network.neutron [None req-121f5830-0aaf-4714-bbd9-ccf079025ad9 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 927.266099] env[61987]: DEBUG nova.objects.instance [None req-121f5830-0aaf-4714-bbd9-ccf079025ad9 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lazy-loading 'info_cache' on Instance uuid 69a22e40-d469-4500-926e-0a12a233f252 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.353132] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061970, 'name': ReconfigVM_Task, 'duration_secs': 0.569368} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.353437] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 876b74ae-cae8-44a9-96ba-e2869ed43be6/876b74ae-cae8-44a9-96ba-e2869ed43be6.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.354074] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-65b7dc60-36a3-4f86-b421-08c3a891724e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.360134] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 927.360134] env[61987]: value = "task-1061971" [ 927.360134] env[61987]: _type = "Task" [ 927.360134] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.371501] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061971, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.472145] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Acquiring lock "cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.472145] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Lock "cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.646018] env[61987]: DEBUG nova.scheduler.client.report [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 927.769401] env[61987]: DEBUG nova.objects.base [None req-121f5830-0aaf-4714-bbd9-ccf079025ad9 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Object Instance<69a22e40-d469-4500-926e-0a12a233f252> lazy-loaded attributes: flavor,info_cache {{(pid=61987) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 927.872111] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061971, 'name': Rename_Task, 'duration_secs': 0.199296} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.873169] env[61987]: DEBUG nova.compute.manager [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 927.875199] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 927.875879] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-da3b5750-04e3-4e25-ad19-d3c6faecab44 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.882710] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 927.882710] env[61987]: value = "task-1061972" [ 927.882710] env[61987]: _type = "Task" [ 927.882710] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.891333] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061972, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.904500] env[61987]: DEBUG nova.virt.hardware [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 927.904895] env[61987]: DEBUG nova.virt.hardware [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 927.905127] env[61987]: DEBUG nova.virt.hardware [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 927.905357] env[61987]: DEBUG nova.virt.hardware [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 927.905568] env[61987]: DEBUG nova.virt.hardware [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 927.905780] env[61987]: DEBUG nova.virt.hardware [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 927.906100] env[61987]: DEBUG nova.virt.hardware [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 927.906287] env[61987]: DEBUG nova.virt.hardware [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 927.906463] env[61987]: DEBUG nova.virt.hardware [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 927.906634] env[61987]: DEBUG nova.virt.hardware [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 927.906811] env[61987]: DEBUG nova.virt.hardware [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 927.908180] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-165f8302-2284-4067-a7a5-4ac628ae0ea5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.915955] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaea7870-f98a-499e-a27f-7db485b4e7fe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.975458] env[61987]: DEBUG nova.compute.manager [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 928.149770] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.308s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.150330] env[61987]: DEBUG nova.compute.manager [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 928.153485] env[61987]: DEBUG oslo_concurrency.lockutils [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.218s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.154972] env[61987]: INFO nova.compute.claims [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 928.258030] env[61987]: DEBUG nova.compute.manager [req-7e71d632-d8ea-4b8a-9f8d-be320656acf1 req-c2291523-0e43-4c2c-bc98-5620164e7f59 service nova] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Received event network-vif-plugged-d857e05f-901f-4f86-9f9e-33aed6d5fec9 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 928.258030] env[61987]: DEBUG oslo_concurrency.lockutils [req-7e71d632-d8ea-4b8a-9f8d-be320656acf1 req-c2291523-0e43-4c2c-bc98-5620164e7f59 service nova] Acquiring lock "ad9041d9-1e0f-46d6-93f7-efe82197deb6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.258030] env[61987]: DEBUG oslo_concurrency.lockutils [req-7e71d632-d8ea-4b8a-9f8d-be320656acf1 req-c2291523-0e43-4c2c-bc98-5620164e7f59 service nova] Lock "ad9041d9-1e0f-46d6-93f7-efe82197deb6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.258030] env[61987]: DEBUG oslo_concurrency.lockutils [req-7e71d632-d8ea-4b8a-9f8d-be320656acf1 req-c2291523-0e43-4c2c-bc98-5620164e7f59 service nova] Lock "ad9041d9-1e0f-46d6-93f7-efe82197deb6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.258030] env[61987]: DEBUG nova.compute.manager [req-7e71d632-d8ea-4b8a-9f8d-be320656acf1 req-c2291523-0e43-4c2c-bc98-5620164e7f59 service nova] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] No waiting events found dispatching network-vif-plugged-d857e05f-901f-4f86-9f9e-33aed6d5fec9 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 928.258030] env[61987]: WARNING nova.compute.manager [req-7e71d632-d8ea-4b8a-9f8d-be320656acf1 req-c2291523-0e43-4c2c-bc98-5620164e7f59 service nova] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Received unexpected event network-vif-plugged-d857e05f-901f-4f86-9f9e-33aed6d5fec9 for instance with vm_state building and task_state spawning. [ 928.393381] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061972, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.420320] env[61987]: DEBUG nova.network.neutron [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Successfully updated port: d857e05f-901f-4f86-9f9e-33aed6d5fec9 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 928.501620] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.563526] env[61987]: DEBUG nova.network.neutron [None req-121f5830-0aaf-4714-bbd9-ccf079025ad9 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updating instance_info_cache with network_info: [{"id": "785edefc-5e84-4a34-97e1-b7c1a08f218e", "address": "fa:16:3e:bf:8b:d7", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap785edefc-5e", "ovs_interfaceid": "785edefc-5e84-4a34-97e1-b7c1a08f218e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.662173] env[61987]: DEBUG nova.compute.utils [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 928.664064] env[61987]: DEBUG nova.compute.manager [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 928.664400] env[61987]: DEBUG nova.network.neutron [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 928.736737] env[61987]: DEBUG nova.policy [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '45ceca37c2624c6aab775e5052a9597a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c7efd6c745f547d3803d5a6184547bda', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 928.893210] env[61987]: DEBUG oslo_vmware.api [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061972, 'name': PowerOnVM_Task, 'duration_secs': 0.973342} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.893619] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 928.893710] env[61987]: INFO nova.compute.manager [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Took 9.16 seconds to spawn the instance on the hypervisor. [ 928.893902] env[61987]: DEBUG nova.compute.manager [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 928.894732] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a7311a2-1843-49b6-96ef-b33520a60b53 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.924367] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "refresh_cache-ad9041d9-1e0f-46d6-93f7-efe82197deb6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.924533] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired lock "refresh_cache-ad9041d9-1e0f-46d6-93f7-efe82197deb6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.924848] env[61987]: DEBUG nova.network.neutron [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 929.032074] env[61987]: DEBUG nova.network.neutron [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Successfully created port: 57c4615c-de74-4af1-8b67-1801365b0ea4 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 929.066339] env[61987]: DEBUG oslo_concurrency.lockutils [None req-121f5830-0aaf-4714-bbd9-ccf079025ad9 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.171234] env[61987]: DEBUG nova.compute.manager [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 929.381834] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5009622a-9c03-400a-8a4a-77a759ee2d31 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.389452] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdbc544f-6c69-47aa-9c6d-1b11bf1eb0ff {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.423221] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ee9cb42-7db5-444f-85f8-7c0d0389d316 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.429479] env[61987]: INFO nova.compute.manager [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Took 29.66 seconds to build instance. [ 929.435886] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d74424ac-3541-4eb3-ab52-0dce40c468ce {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.449378] env[61987]: DEBUG nova.compute.provider_tree [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.457615] env[61987]: DEBUG nova.network.neutron [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 929.594305] env[61987]: DEBUG nova.network.neutron [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Updating instance_info_cache with network_info: [{"id": "d857e05f-901f-4f86-9f9e-33aed6d5fec9", "address": "fa:16:3e:c4:73:1e", "network": {"id": "59d8d7d6-fd71-4dec-802e-e0f9e670ff70", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1845227031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "315a2a3450024ed38e27cd7758f392f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd857e05f-90", "ovs_interfaceid": "d857e05f-901f-4f86-9f9e-33aed6d5fec9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.931591] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8d95be82-782c-4971-9746-a40f698d48f9 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "876b74ae-cae8-44a9-96ba-e2869ed43be6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.174s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.952831] env[61987]: DEBUG nova.scheduler.client.report [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 930.071887] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-121f5830-0aaf-4714-bbd9-ccf079025ad9 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 930.072536] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-edaea885-46e3-45b4-8f77-fe14d868266c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.080638] env[61987]: DEBUG oslo_vmware.api [None req-121f5830-0aaf-4714-bbd9-ccf079025ad9 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 930.080638] env[61987]: value = "task-1061973" [ 930.080638] env[61987]: _type = "Task" [ 930.080638] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.088552] env[61987]: DEBUG oslo_vmware.api [None req-121f5830-0aaf-4714-bbd9-ccf079025ad9 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1061973, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.097175] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Releasing lock "refresh_cache-ad9041d9-1e0f-46d6-93f7-efe82197deb6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.097413] env[61987]: DEBUG nova.compute.manager [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Instance network_info: |[{"id": "d857e05f-901f-4f86-9f9e-33aed6d5fec9", "address": "fa:16:3e:c4:73:1e", "network": {"id": "59d8d7d6-fd71-4dec-802e-e0f9e670ff70", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1845227031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "315a2a3450024ed38e27cd7758f392f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd857e05f-90", "ovs_interfaceid": "d857e05f-901f-4f86-9f9e-33aed6d5fec9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 930.097805] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:73:1e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '028bae2d-fe6c-4207-b4a3-3fab45fbf1d6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd857e05f-901f-4f86-9f9e-33aed6d5fec9', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 930.107887] env[61987]: DEBUG oslo.service.loopingcall [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.108418] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 930.108671] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6992106e-8040-4d23-a548-59f27ee9b7ba {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.128523] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 930.128523] env[61987]: value = "task-1061974" [ 930.128523] env[61987]: _type = "Task" [ 930.128523] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.135842] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061974, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.184503] env[61987]: DEBUG nova.compute.manager [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 930.211949] env[61987]: DEBUG nova.virt.hardware [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 930.211949] env[61987]: DEBUG nova.virt.hardware [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 930.212217] env[61987]: DEBUG nova.virt.hardware [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 930.212277] env[61987]: DEBUG nova.virt.hardware [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 930.212439] env[61987]: DEBUG nova.virt.hardware [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 930.212574] env[61987]: DEBUG nova.virt.hardware [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 930.212823] env[61987]: DEBUG nova.virt.hardware [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 930.213024] env[61987]: DEBUG nova.virt.hardware [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 930.213211] env[61987]: DEBUG nova.virt.hardware [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 930.213386] env[61987]: DEBUG nova.virt.hardware [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 930.213569] env[61987]: DEBUG nova.virt.hardware [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 930.214459] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b832f5de-fed4-4bf6-bf29-71b29c1bc348 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.223168] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f12fe2-cd21-461e-b72e-8811cf76df54 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.285665] env[61987]: DEBUG nova.compute.manager [req-69c93883-6037-406f-b608-12d2734b101d req-684472a1-0046-463d-be3a-4ad98be9c603 service nova] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Received event network-changed-d857e05f-901f-4f86-9f9e-33aed6d5fec9 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 930.285981] env[61987]: DEBUG nova.compute.manager [req-69c93883-6037-406f-b608-12d2734b101d req-684472a1-0046-463d-be3a-4ad98be9c603 service nova] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Refreshing instance network info cache due to event network-changed-d857e05f-901f-4f86-9f9e-33aed6d5fec9. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 930.286272] env[61987]: DEBUG oslo_concurrency.lockutils [req-69c93883-6037-406f-b608-12d2734b101d req-684472a1-0046-463d-be3a-4ad98be9c603 service nova] Acquiring lock "refresh_cache-ad9041d9-1e0f-46d6-93f7-efe82197deb6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.286484] env[61987]: DEBUG oslo_concurrency.lockutils [req-69c93883-6037-406f-b608-12d2734b101d req-684472a1-0046-463d-be3a-4ad98be9c603 service nova] Acquired lock "refresh_cache-ad9041d9-1e0f-46d6-93f7-efe82197deb6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.286647] env[61987]: DEBUG nova.network.neutron [req-69c93883-6037-406f-b608-12d2734b101d req-684472a1-0046-463d-be3a-4ad98be9c603 service nova] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Refreshing network info cache for port d857e05f-901f-4f86-9f9e-33aed6d5fec9 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 930.459523] env[61987]: DEBUG oslo_concurrency.lockutils [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.305s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.459849] env[61987]: DEBUG nova.compute.manager [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 930.462502] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.145s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.462738] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.465441] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.505s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.465690] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.468084] env[61987]: DEBUG oslo_concurrency.lockutils [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.542s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.468084] env[61987]: DEBUG nova.objects.instance [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lazy-loading 'resources' on Instance uuid 51a36102-795e-47b7-b96a-857e54dc703e {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 930.492837] env[61987]: INFO nova.scheduler.client.report [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Deleted allocations for instance f937a5ec-a1d7-41d8-b998-fa18f545b304 [ 930.494431] env[61987]: INFO nova.scheduler.client.report [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Deleted allocations for instance d8b9e172-c3bf-47d9-969b-7f33475d2d60 [ 930.552654] env[61987]: DEBUG nova.network.neutron [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Successfully updated port: 57c4615c-de74-4af1-8b67-1801365b0ea4 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 930.592603] env[61987]: DEBUG oslo_vmware.api [None req-121f5830-0aaf-4714-bbd9-ccf079025ad9 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1061973, 'name': PowerOnVM_Task, 'duration_secs': 0.35545} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.593172] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-121f5830-0aaf-4714-bbd9-ccf079025ad9 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 930.594342] env[61987]: DEBUG nova.compute.manager [None req-121f5830-0aaf-4714-bbd9-ccf079025ad9 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 930.595012] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30c7f0ea-2ff3-4957-9399-6be3e2340bc2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.639351] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061974, 'name': CreateVM_Task, 'duration_secs': 0.304365} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.639556] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 930.640366] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.640568] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.640997] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 930.641189] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b9cf0f2-8931-4190-b62d-ecc2f15198af {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.646154] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 930.646154] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]524e9736-527d-9f29-c392-d1733a4f7e79" [ 930.646154] env[61987]: _type = "Task" [ 930.646154] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.663870] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524e9736-527d-9f29-c392-d1733a4f7e79, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.971509] env[61987]: DEBUG nova.compute.utils [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 930.981940] env[61987]: DEBUG nova.compute.manager [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 930.981940] env[61987]: DEBUG nova.network.neutron [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 931.007617] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e6569c67-9bef-4c0f-a24c-d478b4150b85 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "f937a5ec-a1d7-41d8-b998-fa18f545b304" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.324s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.011033] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9eac87c6-945e-4aad-9fb3-9616b8902cb0 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "d8b9e172-c3bf-47d9-969b-7f33475d2d60" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.720s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.055063] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "refresh_cache-ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.055236] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquired lock "refresh_cache-ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.055392] env[61987]: DEBUG nova.network.neutron [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 931.065344] env[61987]: DEBUG nova.policy [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c124c6b758543a68e141796b585e41f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bd47eb44bc334bf3ae5813905903ecbc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 931.159410] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524e9736-527d-9f29-c392-d1733a4f7e79, 'name': SearchDatastore_Task, 'duration_secs': 0.02017} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.160347] env[61987]: DEBUG nova.network.neutron [req-69c93883-6037-406f-b608-12d2734b101d req-684472a1-0046-463d-be3a-4ad98be9c603 service nova] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Updated VIF entry in instance network info cache for port d857e05f-901f-4f86-9f9e-33aed6d5fec9. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 931.160713] env[61987]: DEBUG nova.network.neutron [req-69c93883-6037-406f-b608-12d2734b101d req-684472a1-0046-463d-be3a-4ad98be9c603 service nova] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Updating instance_info_cache with network_info: [{"id": "d857e05f-901f-4f86-9f9e-33aed6d5fec9", "address": "fa:16:3e:c4:73:1e", "network": {"id": "59d8d7d6-fd71-4dec-802e-e0f9e670ff70", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1845227031-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "315a2a3450024ed38e27cd7758f392f9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "028bae2d-fe6c-4207-b4a3-3fab45fbf1d6", "external-id": "nsx-vlan-transportzone-955", "segmentation_id": 955, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd857e05f-90", "ovs_interfaceid": "d857e05f-901f-4f86-9f9e-33aed6d5fec9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.162483] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.162764] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 931.163173] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.163261] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.163393] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 931.167128] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d41f8dcd-ac35-4f72-91fb-56fffe4aa86e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.178042] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 931.178042] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 931.178042] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f310de7-5cee-43c1-93ba-bc3d6d8a184f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.190022] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 931.190022] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52b1c6bf-6aa4-f5db-0a18-8ff5b35e88d0" [ 931.190022] env[61987]: _type = "Task" [ 931.190022] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.199869] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b1c6bf-6aa4-f5db-0a18-8ff5b35e88d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.243197] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a882a1a4-43bc-43b1-8d67-def50f1d3e60 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.251627] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d54a1208-a06b-449e-aa1c-f1186b1176fb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.284896] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb838b5-ef2d-4ed3-9cbb-26bd02203c2a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.295141] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f547bf60-e244-41cd-9fb4-c49ad8bb159f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.309171] env[61987]: DEBUG nova.compute.provider_tree [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.431711] env[61987]: DEBUG nova.network.neutron [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Successfully created port: f0bccd00-ae35-45ee-af83-b29875006ee1 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 931.482200] env[61987]: DEBUG nova.compute.manager [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 931.624693] env[61987]: DEBUG nova.network.neutron [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 931.664310] env[61987]: DEBUG oslo_concurrency.lockutils [req-69c93883-6037-406f-b608-12d2734b101d req-684472a1-0046-463d-be3a-4ad98be9c603 service nova] Releasing lock "refresh_cache-ad9041d9-1e0f-46d6-93f7-efe82197deb6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.703488] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b1c6bf-6aa4-f5db-0a18-8ff5b35e88d0, 'name': SearchDatastore_Task, 'duration_secs': 0.019975} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.704400] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aef8a619-076a-4fdb-9edc-dc04082f68f8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.710602] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 931.710602] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5206fdfe-ccd0-be0e-61a9-b4eb258f6423" [ 931.710602] env[61987]: _type = "Task" [ 931.710602] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.720865] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5206fdfe-ccd0-be0e-61a9-b4eb258f6423, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.812673] env[61987]: DEBUG nova.scheduler.client.report [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 932.150607] env[61987]: DEBUG nova.network.neutron [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Updating instance_info_cache with network_info: [{"id": "57c4615c-de74-4af1-8b67-1801365b0ea4", "address": "fa:16:3e:73:38:cd", "network": {"id": "a9093c43-6bc4-45b1-bdcf-a434759300bd", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-355385599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7efd6c745f547d3803d5a6184547bda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac563aa7-6d7c-4bd5-9241-7b3e11b8f22d", "external-id": "nsx-vlan-transportzone-730", "segmentation_id": 730, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57c4615c-de", "ovs_interfaceid": "57c4615c-de74-4af1-8b67-1801365b0ea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.226241] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5206fdfe-ccd0-be0e-61a9-b4eb258f6423, 'name': SearchDatastore_Task, 'duration_secs': 0.011234} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.226481] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.226780] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] ad9041d9-1e0f-46d6-93f7-efe82197deb6/ad9041d9-1e0f-46d6-93f7-efe82197deb6.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 932.227106] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-74fd7de7-f944-43ef-b862-b5a55e13ce58 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.234506] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 932.234506] env[61987]: value = "task-1061975" [ 932.234506] env[61987]: _type = "Task" [ 932.234506] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.243499] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061975, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.318507] env[61987]: DEBUG oslo_concurrency.lockutils [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.851s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.321417] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.528s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.323734] env[61987]: INFO nova.compute.claims [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 932.365525] env[61987]: INFO nova.scheduler.client.report [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Deleted allocations for instance 51a36102-795e-47b7-b96a-857e54dc703e [ 932.496590] env[61987]: DEBUG nova.compute.manager [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 932.525192] env[61987]: DEBUG nova.virt.hardware [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 932.525192] env[61987]: DEBUG nova.virt.hardware [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 932.525192] env[61987]: DEBUG nova.virt.hardware [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 932.525192] env[61987]: DEBUG nova.virt.hardware [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 932.525504] env[61987]: DEBUG nova.virt.hardware [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 932.525904] env[61987]: DEBUG nova.virt.hardware [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 932.526324] env[61987]: DEBUG nova.virt.hardware [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 932.526666] env[61987]: DEBUG nova.virt.hardware [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 932.527049] env[61987]: DEBUG nova.virt.hardware [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 932.527419] env[61987]: DEBUG nova.virt.hardware [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 932.530134] env[61987]: DEBUG nova.virt.hardware [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 932.530134] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f857918-fd3d-41dc-a1eb-493889ffe937 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.539722] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-532940d5-49aa-4fd5-a169-c6b6ddb34381 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.658029] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Releasing lock "refresh_cache-ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.658029] env[61987]: DEBUG nova.compute.manager [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Instance network_info: |[{"id": "57c4615c-de74-4af1-8b67-1801365b0ea4", "address": "fa:16:3e:73:38:cd", "network": {"id": "a9093c43-6bc4-45b1-bdcf-a434759300bd", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-355385599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7efd6c745f547d3803d5a6184547bda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac563aa7-6d7c-4bd5-9241-7b3e11b8f22d", "external-id": "nsx-vlan-transportzone-730", "segmentation_id": 730, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57c4615c-de", "ovs_interfaceid": "57c4615c-de74-4af1-8b67-1801365b0ea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 932.658029] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:73:38:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ac563aa7-6d7c-4bd5-9241-7b3e11b8f22d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '57c4615c-de74-4af1-8b67-1801365b0ea4', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 932.664239] env[61987]: DEBUG oslo.service.loopingcall [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 932.664833] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 932.665303] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f1385ee2-7ecb-46d3-b500-fea6728cc6ae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.689602] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 932.689602] env[61987]: value = "task-1061976" [ 932.689602] env[61987]: _type = "Task" [ 932.689602] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.709229] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061976, 'name': CreateVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.750030] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061975, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.507339} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.750030] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] ad9041d9-1e0f-46d6-93f7-efe82197deb6/ad9041d9-1e0f-46d6-93f7-efe82197deb6.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 932.750030] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 932.750030] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8c19f150-532b-4fc1-95ac-a6a16c02a0bd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.759020] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 932.759020] env[61987]: value = "task-1061977" [ 932.759020] env[61987]: _type = "Task" [ 932.759020] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.767770] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061977, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.870161] env[61987]: DEBUG nova.compute.manager [req-2735d0eb-6565-444a-97c5-3929f0add5de req-d8a29eff-b8ef-4b3d-8941-ac588834c208 service nova] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Received event network-vif-plugged-57c4615c-de74-4af1-8b67-1801365b0ea4 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 932.870384] env[61987]: DEBUG oslo_concurrency.lockutils [req-2735d0eb-6565-444a-97c5-3929f0add5de req-d8a29eff-b8ef-4b3d-8941-ac588834c208 service nova] Acquiring lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.870603] env[61987]: DEBUG oslo_concurrency.lockutils [req-2735d0eb-6565-444a-97c5-3929f0add5de req-d8a29eff-b8ef-4b3d-8941-ac588834c208 service nova] Lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.870763] env[61987]: DEBUG oslo_concurrency.lockutils [req-2735d0eb-6565-444a-97c5-3929f0add5de req-d8a29eff-b8ef-4b3d-8941-ac588834c208 service nova] Lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.870936] env[61987]: DEBUG nova.compute.manager [req-2735d0eb-6565-444a-97c5-3929f0add5de req-d8a29eff-b8ef-4b3d-8941-ac588834c208 service nova] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] No waiting events found dispatching network-vif-plugged-57c4615c-de74-4af1-8b67-1801365b0ea4 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 932.871358] env[61987]: WARNING nova.compute.manager [req-2735d0eb-6565-444a-97c5-3929f0add5de req-d8a29eff-b8ef-4b3d-8941-ac588834c208 service nova] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Received unexpected event network-vif-plugged-57c4615c-de74-4af1-8b67-1801365b0ea4 for instance with vm_state building and task_state spawning. [ 932.871643] env[61987]: DEBUG nova.compute.manager [req-2735d0eb-6565-444a-97c5-3929f0add5de req-d8a29eff-b8ef-4b3d-8941-ac588834c208 service nova] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Received event network-changed-57c4615c-de74-4af1-8b67-1801365b0ea4 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 932.871822] env[61987]: DEBUG nova.compute.manager [req-2735d0eb-6565-444a-97c5-3929f0add5de req-d8a29eff-b8ef-4b3d-8941-ac588834c208 service nova] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Refreshing instance network info cache due to event network-changed-57c4615c-de74-4af1-8b67-1801365b0ea4. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 932.872097] env[61987]: DEBUG oslo_concurrency.lockutils [req-2735d0eb-6565-444a-97c5-3929f0add5de req-d8a29eff-b8ef-4b3d-8941-ac588834c208 service nova] Acquiring lock "refresh_cache-ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.872302] env[61987]: DEBUG oslo_concurrency.lockutils [req-2735d0eb-6565-444a-97c5-3929f0add5de req-d8a29eff-b8ef-4b3d-8941-ac588834c208 service nova] Acquired lock "refresh_cache-ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.872685] env[61987]: DEBUG nova.network.neutron [req-2735d0eb-6565-444a-97c5-3929f0add5de req-d8a29eff-b8ef-4b3d-8941-ac588834c208 service nova] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Refreshing network info cache for port 57c4615c-de74-4af1-8b67-1801365b0ea4 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 932.879283] env[61987]: DEBUG oslo_concurrency.lockutils [None req-192aa65f-38ff-4b46-9a34-7c70c32bb9ad tempest-ServersNegativeTestJSON-2100701923 tempest-ServersNegativeTestJSON-2100701923-project-member] Lock "51a36102-795e-47b7-b96a-857e54dc703e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.815s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.201781] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061976, 'name': CreateVM_Task, 'duration_secs': 0.40989} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.202051] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 933.202689] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.202880] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.203224] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 933.203485] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66fdaf9b-4a9e-4a51-831e-6c351e3d98b7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.211519] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 933.211519] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52719c97-8c35-025b-6f8a-8b6882ff67b1" [ 933.211519] env[61987]: _type = "Task" [ 933.211519] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.219886] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52719c97-8c35-025b-6f8a-8b6882ff67b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.269376] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061977, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065997} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.269680] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 933.270719] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ece5b26-56f1-4d54-a79a-fb0661a74631 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.296109] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] ad9041d9-1e0f-46d6-93f7-efe82197deb6/ad9041d9-1e0f-46d6-93f7-efe82197deb6.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 933.296477] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c35f5328-11c3-4bf1-aa8c-fd9b1f0759b8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.320557] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 933.320557] env[61987]: value = "task-1061978" [ 933.320557] env[61987]: _type = "Task" [ 933.320557] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.336219] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061978, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.427121] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6857a837-31a5-427e-8fbc-66d37dd69d03 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.427592] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6857a837-31a5-427e-8fbc-66d37dd69d03 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.427973] env[61987]: DEBUG nova.compute.manager [None req-6857a837-31a5-427e-8fbc-66d37dd69d03 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 933.429481] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcabdb01-d63d-4fa0-86c6-ee13565603b1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.437371] env[61987]: DEBUG nova.compute.manager [None req-6857a837-31a5-427e-8fbc-66d37dd69d03 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61987) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 933.437631] env[61987]: DEBUG nova.objects.instance [None req-6857a837-31a5-427e-8fbc-66d37dd69d03 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lazy-loading 'flavor' on Instance uuid 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 933.546121] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1526f099-fa2e-4490-a1dd-ca22c7d04bb1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.552667] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44e73eca-bcb8-4745-9837-949167db302e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.591403] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ee7d820-945a-47d6-a3ff-f0161ab1f257 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.601407] env[61987]: DEBUG nova.network.neutron [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Successfully updated port: f0bccd00-ae35-45ee-af83-b29875006ee1 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 933.606091] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4731820-bffe-4a6f-9e66-159fdf4bee61 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.625973] env[61987]: DEBUG nova.compute.provider_tree [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.721546] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52719c97-8c35-025b-6f8a-8b6882ff67b1, 'name': SearchDatastore_Task, 'duration_secs': 0.069706} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.722171] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.722465] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 933.722784] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.722944] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.723179] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 933.726084] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-50581e76-b182-43c6-8967-5aeac2edb1a9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.737713] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 933.737913] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 933.738679] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f33c78f7-00b0-4a4d-afdd-0d78711c442a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.747621] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 933.747621] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5257eb1e-c621-0e7a-065b-05ca76358b13" [ 933.747621] env[61987]: _type = "Task" [ 933.747621] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.753249] env[61987]: DEBUG nova.compute.manager [req-c66ef81a-aeee-4e78-bf4c-f5304ce0da13 req-4eea4977-0e0e-473a-af9c-e77bec72da72 service nova] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Received event network-vif-plugged-f0bccd00-ae35-45ee-af83-b29875006ee1 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 933.753510] env[61987]: DEBUG oslo_concurrency.lockutils [req-c66ef81a-aeee-4e78-bf4c-f5304ce0da13 req-4eea4977-0e0e-473a-af9c-e77bec72da72 service nova] Acquiring lock "a720083b-7b16-468a-a595-bfe8592e8621-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.753923] env[61987]: DEBUG oslo_concurrency.lockutils [req-c66ef81a-aeee-4e78-bf4c-f5304ce0da13 req-4eea4977-0e0e-473a-af9c-e77bec72da72 service nova] Lock "a720083b-7b16-468a-a595-bfe8592e8621-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.754086] env[61987]: DEBUG oslo_concurrency.lockutils [req-c66ef81a-aeee-4e78-bf4c-f5304ce0da13 req-4eea4977-0e0e-473a-af9c-e77bec72da72 service nova] Lock "a720083b-7b16-468a-a595-bfe8592e8621-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.754339] env[61987]: DEBUG nova.compute.manager [req-c66ef81a-aeee-4e78-bf4c-f5304ce0da13 req-4eea4977-0e0e-473a-af9c-e77bec72da72 service nova] [instance: a720083b-7b16-468a-a595-bfe8592e8621] No waiting events found dispatching network-vif-plugged-f0bccd00-ae35-45ee-af83-b29875006ee1 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 933.754608] env[61987]: WARNING nova.compute.manager [req-c66ef81a-aeee-4e78-bf4c-f5304ce0da13 req-4eea4977-0e0e-473a-af9c-e77bec72da72 service nova] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Received unexpected event network-vif-plugged-f0bccd00-ae35-45ee-af83-b29875006ee1 for instance with vm_state building and task_state spawning. [ 933.761449] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5257eb1e-c621-0e7a-065b-05ca76358b13, 'name': SearchDatastore_Task, 'duration_secs': 0.00953} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.762589] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6408555-bc7e-4eeb-a61f-ea9ea0a1b6aa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.767832] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 933.767832] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52c98ca5-f6fc-e8f8-0ae1-109c91849bff" [ 933.767832] env[61987]: _type = "Task" [ 933.767832] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.776714] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c98ca5-f6fc-e8f8-0ae1-109c91849bff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.816998] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquiring lock "455562c7-9fc1-4625-95b0-5eb8c9c8ccaa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.817360] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "455562c7-9fc1-4625-95b0-5eb8c9c8ccaa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.817610] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquiring lock "455562c7-9fc1-4625-95b0-5eb8c9c8ccaa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.817973] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "455562c7-9fc1-4625-95b0-5eb8c9c8ccaa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.818381] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "455562c7-9fc1-4625-95b0-5eb8c9c8ccaa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.821298] env[61987]: INFO nova.compute.manager [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Terminating instance [ 933.837194] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061978, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.897483] env[61987]: DEBUG nova.network.neutron [req-2735d0eb-6565-444a-97c5-3929f0add5de req-d8a29eff-b8ef-4b3d-8941-ac588834c208 service nova] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Updated VIF entry in instance network info cache for port 57c4615c-de74-4af1-8b67-1801365b0ea4. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 933.897483] env[61987]: DEBUG nova.network.neutron [req-2735d0eb-6565-444a-97c5-3929f0add5de req-d8a29eff-b8ef-4b3d-8941-ac588834c208 service nova] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Updating instance_info_cache with network_info: [{"id": "57c4615c-de74-4af1-8b67-1801365b0ea4", "address": "fa:16:3e:73:38:cd", "network": {"id": "a9093c43-6bc4-45b1-bdcf-a434759300bd", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-355385599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7efd6c745f547d3803d5a6184547bda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac563aa7-6d7c-4bd5-9241-7b3e11b8f22d", "external-id": "nsx-vlan-transportzone-730", "segmentation_id": 730, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57c4615c-de", "ovs_interfaceid": "57c4615c-de74-4af1-8b67-1801365b0ea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.105572] env[61987]: DEBUG oslo_concurrency.lockutils [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "refresh_cache-a720083b-7b16-468a-a595-bfe8592e8621" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.105870] env[61987]: DEBUG oslo_concurrency.lockutils [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired lock "refresh_cache-a720083b-7b16-468a-a595-bfe8592e8621" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.106014] env[61987]: DEBUG nova.network.neutron [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 934.132903] env[61987]: DEBUG nova.scheduler.client.report [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 934.278215] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c98ca5-f6fc-e8f8-0ae1-109c91849bff, 'name': SearchDatastore_Task, 'duration_secs': 0.008705} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.278567] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.278835] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] ae1396b3-c28b-4d1c-9a3a-3dc2170847ca/ae1396b3-c28b-4d1c-9a3a-3dc2170847ca.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 934.279431] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-714c5cd3-3803-42ba-a200-52a52d3c2cd5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.285485] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 934.285485] env[61987]: value = "task-1061979" [ 934.285485] env[61987]: _type = "Task" [ 934.285485] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.293023] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061979, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.333101] env[61987]: DEBUG nova.compute.manager [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 934.333354] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 934.333655] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061978, 'name': ReconfigVM_Task, 'duration_secs': 0.632919} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.334387] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beb6b8c5-47f0-4724-accc-cd8ce1805de4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.337336] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Reconfigured VM instance instance-00000059 to attach disk [datastore2] ad9041d9-1e0f-46d6-93f7-efe82197deb6/ad9041d9-1e0f-46d6-93f7-efe82197deb6.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 934.337810] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-73831275-3d51-4453-90b9-30ba0ef46b69 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.343562] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 934.344688] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-244ef87c-0877-4af3-8d0a-e6ca38a77540 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.346154] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 934.346154] env[61987]: value = "task-1061980" [ 934.346154] env[61987]: _type = "Task" [ 934.346154] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.351024] env[61987]: DEBUG oslo_vmware.api [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for the task: (returnval){ [ 934.351024] env[61987]: value = "task-1061981" [ 934.351024] env[61987]: _type = "Task" [ 934.351024] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.362984] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061980, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.371674] env[61987]: DEBUG oslo_vmware.api [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061981, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.400616] env[61987]: DEBUG oslo_concurrency.lockutils [req-2735d0eb-6565-444a-97c5-3929f0add5de req-d8a29eff-b8ef-4b3d-8941-ac588834c208 service nova] Releasing lock "refresh_cache-ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.453045] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6857a837-31a5-427e-8fbc-66d37dd69d03 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 934.453045] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5dc09cfd-79fc-4e4e-bd5f-3ab7b2964473 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.458378] env[61987]: DEBUG oslo_vmware.api [None req-6857a837-31a5-427e-8fbc-66d37dd69d03 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 934.458378] env[61987]: value = "task-1061982" [ 934.458378] env[61987]: _type = "Task" [ 934.458378] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.468254] env[61987]: DEBUG oslo_vmware.api [None req-6857a837-31a5-427e-8fbc-66d37dd69d03 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061982, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.535525] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "865222ae-12da-4e5a-84af-041003b50520" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.535525] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "865222ae-12da-4e5a-84af-041003b50520" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.641319] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.317s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.641319] env[61987]: DEBUG nova.compute.manager [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 934.643672] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.142s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.648369] env[61987]: INFO nova.compute.claims [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 934.685728] env[61987]: DEBUG nova.network.neutron [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 934.799778] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061979, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.860973] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061980, 'name': Rename_Task, 'duration_secs': 0.165448} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.864484] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 934.864897] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d5724268-e7ea-4df5-8015-2c2107a9521b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.866313] env[61987]: DEBUG oslo_vmware.api [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061981, 'name': PowerOffVM_Task, 'duration_secs': 0.375513} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.866613] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 934.866836] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 934.867398] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-626407da-cca3-4254-8df5-67fd5e6d64ee {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.871605] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 934.871605] env[61987]: value = "task-1061983" [ 934.871605] env[61987]: _type = "Task" [ 934.871605] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.887432] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061983, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.907449] env[61987]: DEBUG nova.compute.manager [req-7cdd8c78-2e85-430e-ab42-c0051ef25f74 req-5d625e77-335f-4e5d-931b-6f416c051922 service nova] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Received event network-changed-f0bccd00-ae35-45ee-af83-b29875006ee1 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 934.909060] env[61987]: DEBUG nova.compute.manager [req-7cdd8c78-2e85-430e-ab42-c0051ef25f74 req-5d625e77-335f-4e5d-931b-6f416c051922 service nova] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Refreshing instance network info cache due to event network-changed-f0bccd00-ae35-45ee-af83-b29875006ee1. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 934.909361] env[61987]: DEBUG oslo_concurrency.lockutils [req-7cdd8c78-2e85-430e-ab42-c0051ef25f74 req-5d625e77-335f-4e5d-931b-6f416c051922 service nova] Acquiring lock "refresh_cache-a720083b-7b16-468a-a595-bfe8592e8621" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.927372] env[61987]: DEBUG nova.network.neutron [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Updating instance_info_cache with network_info: [{"id": "f0bccd00-ae35-45ee-af83-b29875006ee1", "address": "fa:16:3e:13:c6:31", "network": {"id": "ca68bf96-3f18-45ee-80fb-64a1260ed996", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-733595377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd47eb44bc334bf3ae5813905903ecbc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a27fd90b-16a5-43af-bede-ae36762ece00", "external-id": "nsx-vlan-transportzone-197", "segmentation_id": 197, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0bccd00-ae", "ovs_interfaceid": "f0bccd00-ae35-45ee-af83-b29875006ee1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.931200] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 934.931677] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 934.931677] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Deleting the datastore file [datastore2] 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 934.931919] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc1c78eb-f49e-4272-a993-3a75afd90904 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.941910] env[61987]: DEBUG oslo_vmware.api [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for the task: (returnval){ [ 934.941910] env[61987]: value = "task-1061985" [ 934.941910] env[61987]: _type = "Task" [ 934.941910] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.952959] env[61987]: DEBUG oslo_vmware.api [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061985, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.971068] env[61987]: DEBUG oslo_vmware.api [None req-6857a837-31a5-427e-8fbc-66d37dd69d03 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061982, 'name': PowerOffVM_Task, 'duration_secs': 0.268271} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.971068] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6857a837-31a5-427e-8fbc-66d37dd69d03 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 934.971068] env[61987]: DEBUG nova.compute.manager [None req-6857a837-31a5-427e-8fbc-66d37dd69d03 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 934.971068] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a9d8826-0812-4c8d-a69c-fbf78d220cdf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.039019] env[61987]: DEBUG nova.compute.manager [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 935.155747] env[61987]: DEBUG nova.compute.utils [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 935.159344] env[61987]: DEBUG nova.compute.manager [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 935.159530] env[61987]: DEBUG nova.network.neutron [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 935.202235] env[61987]: DEBUG nova.policy [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e16f557dfdb04406b3fc772b891f05a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '22e62741abfb4471bbc7e11ae7d407ca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 935.296982] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061979, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.549024} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.296982] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] ae1396b3-c28b-4d1c-9a3a-3dc2170847ca/ae1396b3-c28b-4d1c-9a3a-3dc2170847ca.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 935.297434] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 935.297434] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8f2e8e85-e812-46e3-b54a-ef5d25908168 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.304110] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 935.304110] env[61987]: value = "task-1061986" [ 935.304110] env[61987]: _type = "Task" [ 935.304110] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.311536] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061986, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.381788] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061983, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.430800] env[61987]: DEBUG oslo_concurrency.lockutils [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Releasing lock "refresh_cache-a720083b-7b16-468a-a595-bfe8592e8621" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.431280] env[61987]: DEBUG nova.compute.manager [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Instance network_info: |[{"id": "f0bccd00-ae35-45ee-af83-b29875006ee1", "address": "fa:16:3e:13:c6:31", "network": {"id": "ca68bf96-3f18-45ee-80fb-64a1260ed996", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-733595377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd47eb44bc334bf3ae5813905903ecbc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a27fd90b-16a5-43af-bede-ae36762ece00", "external-id": "nsx-vlan-transportzone-197", "segmentation_id": 197, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0bccd00-ae", "ovs_interfaceid": "f0bccd00-ae35-45ee-af83-b29875006ee1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 935.431901] env[61987]: DEBUG oslo_concurrency.lockutils [req-7cdd8c78-2e85-430e-ab42-c0051ef25f74 req-5d625e77-335f-4e5d-931b-6f416c051922 service nova] Acquired lock "refresh_cache-a720083b-7b16-468a-a595-bfe8592e8621" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.432119] env[61987]: DEBUG nova.network.neutron [req-7cdd8c78-2e85-430e-ab42-c0051ef25f74 req-5d625e77-335f-4e5d-931b-6f416c051922 service nova] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Refreshing network info cache for port f0bccd00-ae35-45ee-af83-b29875006ee1 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 935.433564] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:13:c6:31', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a27fd90b-16a5-43af-bede-ae36762ece00', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f0bccd00-ae35-45ee-af83-b29875006ee1', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 935.442623] env[61987]: DEBUG oslo.service.loopingcall [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.444990] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 935.448357] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-525f3d9b-3e94-4625-a4b5-740066355116 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.463628] env[61987]: DEBUG nova.network.neutron [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Successfully created port: 114c5cda-5986-410d-a9ef-59f54ab06c10 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 935.471997] env[61987]: DEBUG oslo_vmware.api [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Task: {'id': task-1061985, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.25453} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.473395] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.473612] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 935.473804] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 935.474065] env[61987]: INFO nova.compute.manager [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Took 1.14 seconds to destroy the instance on the hypervisor. [ 935.474242] env[61987]: DEBUG oslo.service.loopingcall [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.474548] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 935.474548] env[61987]: value = "task-1061987" [ 935.474548] env[61987]: _type = "Task" [ 935.474548] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.474743] env[61987]: DEBUG nova.compute.manager [-] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 935.474743] env[61987]: DEBUG nova.network.neutron [-] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 935.487029] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6857a837-31a5-427e-8fbc-66d37dd69d03 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.059s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.494281] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061987, 'name': CreateVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.567146] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.659663] env[61987]: DEBUG nova.compute.manager [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 935.787154] env[61987]: DEBUG nova.compute.manager [req-1ba068f2-c07f-4f36-94b0-d10ff097e5fa req-b4131503-0365-41e0-961c-dcc31f218bf3 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Received event network-vif-deleted-e4942ff0-6c8c-4c29-850d-778e6caa8e44 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 935.787450] env[61987]: INFO nova.compute.manager [req-1ba068f2-c07f-4f36-94b0-d10ff097e5fa req-b4131503-0365-41e0-961c-dcc31f218bf3 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Neutron deleted interface e4942ff0-6c8c-4c29-850d-778e6caa8e44; detaching it from the instance and deleting it from the info cache [ 935.787569] env[61987]: DEBUG nova.network.neutron [req-1ba068f2-c07f-4f36-94b0-d10ff097e5fa req-b4131503-0365-41e0-961c-dcc31f218bf3 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.817733] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061986, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069123} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.818092] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 935.819133] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9abbc6d-dc48-4614-bb8b-522b1280c0fc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.850219] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] ae1396b3-c28b-4d1c-9a3a-3dc2170847ca/ae1396b3-c28b-4d1c-9a3a-3dc2170847ca.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 935.851937] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-65af06e4-7d9e-4f7a-bb0a-a22a588eca92 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.878322] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 935.878322] env[61987]: value = "task-1061988" [ 935.878322] env[61987]: _type = "Task" [ 935.878322] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.890199] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061983, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.897692] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061988, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.929169] env[61987]: DEBUG nova.network.neutron [req-7cdd8c78-2e85-430e-ab42-c0051ef25f74 req-5d625e77-335f-4e5d-931b-6f416c051922 service nova] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Updated VIF entry in instance network info cache for port f0bccd00-ae35-45ee-af83-b29875006ee1. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 935.929680] env[61987]: DEBUG nova.network.neutron [req-7cdd8c78-2e85-430e-ab42-c0051ef25f74 req-5d625e77-335f-4e5d-931b-6f416c051922 service nova] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Updating instance_info_cache with network_info: [{"id": "f0bccd00-ae35-45ee-af83-b29875006ee1", "address": "fa:16:3e:13:c6:31", "network": {"id": "ca68bf96-3f18-45ee-80fb-64a1260ed996", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-733595377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd47eb44bc334bf3ae5813905903ecbc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a27fd90b-16a5-43af-bede-ae36762ece00", "external-id": "nsx-vlan-transportzone-197", "segmentation_id": 197, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0bccd00-ae", "ovs_interfaceid": "f0bccd00-ae35-45ee-af83-b29875006ee1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.985991] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77d212cd-99b1-497f-99de-257d5bba21ab {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.989246] env[61987]: DEBUG nova.objects.instance [None req-5f9e10f8-d821-4c8d-b8e9-7d9400babbed tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lazy-loading 'flavor' on Instance uuid 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.996579] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061987, 'name': CreateVM_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.999877] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d307f3-29b8-4e06-b37f-d6df1987e7cc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.034500] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5375f0bb-2701-4afd-86ed-17dae6b6bb19 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.042815] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d682fe6e-a00c-49ef-b51c-9baadb3f2976 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.058186] env[61987]: DEBUG nova.compute.provider_tree [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.256993] env[61987]: DEBUG nova.network.neutron [-] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.289679] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d447740c-4594-4703-af23-8c7a69168397 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.298671] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da04473f-b7fa-407f-ba86-41235c3242f0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.330946] env[61987]: DEBUG nova.compute.manager [req-1ba068f2-c07f-4f36-94b0-d10ff097e5fa req-b4131503-0365-41e0-961c-dcc31f218bf3 service nova] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Detach interface failed, port_id=e4942ff0-6c8c-4c29-850d-778e6caa8e44, reason: Instance 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 936.388625] env[61987]: DEBUG oslo_vmware.api [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061983, 'name': PowerOnVM_Task, 'duration_secs': 1.168004} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.389781] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 936.389990] env[61987]: INFO nova.compute.manager [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Took 8.52 seconds to spawn the instance on the hypervisor. [ 936.390291] env[61987]: DEBUG nova.compute.manager [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 936.391544] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61467581-fb1f-4fbd-9ee9-b47aaf77e44e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.399727] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061988, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.435605] env[61987]: DEBUG oslo_concurrency.lockutils [req-7cdd8c78-2e85-430e-ab42-c0051ef25f74 req-5d625e77-335f-4e5d-931b-6f416c051922 service nova] Releasing lock "refresh_cache-a720083b-7b16-468a-a595-bfe8592e8621" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.487948] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061987, 'name': CreateVM_Task, 'duration_secs': 0.784902} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.488147] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 936.488853] env[61987]: DEBUG oslo_concurrency.lockutils [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.489046] env[61987]: DEBUG oslo_concurrency.lockutils [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.489391] env[61987]: DEBUG oslo_concurrency.lockutils [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 936.489653] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3cb17e20-a5db-4fbc-86b7-4e006a5d31a3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.496027] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 936.496027] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5241b2a0-3000-dd0a-b882-a7a1219a0003" [ 936.496027] env[61987]: _type = "Task" [ 936.496027] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.496536] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5f9e10f8-d821-4c8d-b8e9-7d9400babbed tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "refresh_cache-5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.496756] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5f9e10f8-d821-4c8d-b8e9-7d9400babbed tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquired lock "refresh_cache-5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.496961] env[61987]: DEBUG nova.network.neutron [None req-5f9e10f8-d821-4c8d-b8e9-7d9400babbed tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 936.497186] env[61987]: DEBUG nova.objects.instance [None req-5f9e10f8-d821-4c8d-b8e9-7d9400babbed tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lazy-loading 'info_cache' on Instance uuid 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 936.508414] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5241b2a0-3000-dd0a-b882-a7a1219a0003, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.561918] env[61987]: DEBUG nova.scheduler.client.report [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 936.674128] env[61987]: DEBUG nova.compute.manager [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 936.702031] env[61987]: DEBUG nova.virt.hardware [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 936.702031] env[61987]: DEBUG nova.virt.hardware [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 936.702031] env[61987]: DEBUG nova.virt.hardware [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 936.702031] env[61987]: DEBUG nova.virt.hardware [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 936.702031] env[61987]: DEBUG nova.virt.hardware [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 936.702031] env[61987]: DEBUG nova.virt.hardware [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 936.702031] env[61987]: DEBUG nova.virt.hardware [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 936.702500] env[61987]: DEBUG nova.virt.hardware [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 936.702856] env[61987]: DEBUG nova.virt.hardware [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 936.703174] env[61987]: DEBUG nova.virt.hardware [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 936.703478] env[61987]: DEBUG nova.virt.hardware [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 936.704538] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a768b97f-3df0-4470-b8a8-936251d1327f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.713473] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e52f7f2-7213-4b60-990e-fe1b64641e4d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.760311] env[61987]: INFO nova.compute.manager [-] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Took 1.29 seconds to deallocate network for instance. [ 936.890885] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061988, 'name': ReconfigVM_Task, 'duration_secs': 0.560166} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.890991] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Reconfigured VM instance instance-0000005a to attach disk [datastore2] ae1396b3-c28b-4d1c-9a3a-3dc2170847ca/ae1396b3-c28b-4d1c-9a3a-3dc2170847ca.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 936.891614] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-40540897-1a9d-4764-9ded-7e82fc738a7b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.897402] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 936.897402] env[61987]: value = "task-1061989" [ 936.897402] env[61987]: _type = "Task" [ 936.897402] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.905338] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061989, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.914592] env[61987]: INFO nova.compute.manager [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Took 33.27 seconds to build instance. [ 937.004246] env[61987]: DEBUG nova.objects.base [None req-5f9e10f8-d821-4c8d-b8e9-7d9400babbed tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Object Instance<5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32> lazy-loaded attributes: flavor,info_cache {{(pid=61987) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 937.012302] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5241b2a0-3000-dd0a-b882-a7a1219a0003, 'name': SearchDatastore_Task, 'duration_secs': 0.014595} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.013237] env[61987]: DEBUG oslo_concurrency.lockutils [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.013743] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 937.013743] env[61987]: DEBUG oslo_concurrency.lockutils [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.013875] env[61987]: DEBUG oslo_concurrency.lockutils [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.014028] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 937.014516] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-87f942ce-9771-4e4c-875e-a190a9224fe6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.022368] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 937.022570] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 937.023301] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e84dc931-86ea-4a32-b317-ad73742c13aa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.028192] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 937.028192] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]525cdd0f-8d48-c7a7-6a5c-cb40c85d3dd5" [ 937.028192] env[61987]: _type = "Task" [ 937.028192] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.036026] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]525cdd0f-8d48-c7a7-6a5c-cb40c85d3dd5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.066575] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.423s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.067119] env[61987]: DEBUG nova.compute.manager [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 937.069732] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.503s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.071171] env[61987]: INFO nova.compute.claims [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 937.245507] env[61987]: DEBUG nova.compute.manager [req-29dfff15-3869-4cd2-b9b9-492081a97b63 req-5020c074-e341-4417-b124-b029571f84fb service nova] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Received event network-vif-plugged-114c5cda-5986-410d-a9ef-59f54ab06c10 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 937.245831] env[61987]: DEBUG oslo_concurrency.lockutils [req-29dfff15-3869-4cd2-b9b9-492081a97b63 req-5020c074-e341-4417-b124-b029571f84fb service nova] Acquiring lock "f9c82f38-7e06-4949-80a5-6eaa3c01c88c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.246106] env[61987]: DEBUG oslo_concurrency.lockutils [req-29dfff15-3869-4cd2-b9b9-492081a97b63 req-5020c074-e341-4417-b124-b029571f84fb service nova] Lock "f9c82f38-7e06-4949-80a5-6eaa3c01c88c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.246519] env[61987]: DEBUG oslo_concurrency.lockutils [req-29dfff15-3869-4cd2-b9b9-492081a97b63 req-5020c074-e341-4417-b124-b029571f84fb service nova] Lock "f9c82f38-7e06-4949-80a5-6eaa3c01c88c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.246796] env[61987]: DEBUG nova.compute.manager [req-29dfff15-3869-4cd2-b9b9-492081a97b63 req-5020c074-e341-4417-b124-b029571f84fb service nova] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] No waiting events found dispatching network-vif-plugged-114c5cda-5986-410d-a9ef-59f54ab06c10 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 937.247097] env[61987]: WARNING nova.compute.manager [req-29dfff15-3869-4cd2-b9b9-492081a97b63 req-5020c074-e341-4417-b124-b029571f84fb service nova] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Received unexpected event network-vif-plugged-114c5cda-5986-410d-a9ef-59f54ab06c10 for instance with vm_state building and task_state spawning. [ 937.267402] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.384999] env[61987]: DEBUG nova.network.neutron [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Successfully updated port: 114c5cda-5986-410d-a9ef-59f54ab06c10 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 937.409534] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061989, 'name': Rename_Task, 'duration_secs': 0.362682} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.409849] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 937.410126] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e77c85eb-b349-4739-8b39-f81ad111e2e0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.418098] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f41f8c71-6fd5-4858-86d6-9bf524447f5a tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "ad9041d9-1e0f-46d6-93f7-efe82197deb6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.792s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.418753] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 937.418753] env[61987]: value = "task-1061990" [ 937.418753] env[61987]: _type = "Task" [ 937.418753] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.427319] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061990, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.539772] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]525cdd0f-8d48-c7a7-6a5c-cb40c85d3dd5, 'name': SearchDatastore_Task, 'duration_secs': 0.011029} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.542647] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a29df02-5d97-4155-9829-afb17be9066c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.547895] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 937.547895] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52dde2a3-2a3b-5606-95f9-8c2fb6121b29" [ 937.547895] env[61987]: _type = "Task" [ 937.547895] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.557092] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52dde2a3-2a3b-5606-95f9-8c2fb6121b29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.578860] env[61987]: DEBUG nova.compute.utils [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 937.580397] env[61987]: DEBUG nova.compute.manager [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 937.580587] env[61987]: DEBUG nova.network.neutron [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 937.862772] env[61987]: DEBUG nova.policy [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a003b173572449d9a77858eb2609339f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e7f68eb92e124eab9beb7b60ddcc1df3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 937.890150] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "refresh_cache-f9c82f38-7e06-4949-80a5-6eaa3c01c88c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.890321] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired lock "refresh_cache-f9c82f38-7e06-4949-80a5-6eaa3c01c88c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.890420] env[61987]: DEBUG nova.network.neutron [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 938.727296] env[61987]: DEBUG nova.compute.manager [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 938.732829] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061990, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.733035] env[61987]: WARNING oslo_vmware.common.loopingcall [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] task run outlasted interval by 0.313636 sec [ 938.756521] env[61987]: DEBUG oslo_vmware.api [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1061990, 'name': PowerOnVM_Task, 'duration_secs': 0.686621} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.756788] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52dde2a3-2a3b-5606-95f9-8c2fb6121b29, 'name': SearchDatastore_Task, 'duration_secs': 0.018723} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.757327] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 938.757542] env[61987]: INFO nova.compute.manager [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Took 8.57 seconds to spawn the instance on the hypervisor. [ 938.757727] env[61987]: DEBUG nova.compute.manager [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 938.758037] env[61987]: DEBUG oslo_concurrency.lockutils [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.758289] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] a720083b-7b16-468a-a595-bfe8592e8621/a720083b-7b16-468a-a595-bfe8592e8621.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 938.759068] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6e19645-27ed-4219-bb5b-54da62d773de {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.761712] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d4696adb-60ee-4069-8e37-2fb01df015a5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.775693] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 938.775693] env[61987]: value = "task-1061991" [ 938.775693] env[61987]: _type = "Task" [ 938.775693] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.786925] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061991, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.791193] env[61987]: DEBUG nova.network.neutron [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 938.872477] env[61987]: DEBUG nova.network.neutron [None req-5f9e10f8-d821-4c8d-b8e9-7d9400babbed tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Updating instance_info_cache with network_info: [{"id": "a60eaa2f-cdd3-4b07-9a40-d48a4dd95d17", "address": "fa:16:3e:8f:7f:02", "network": {"id": "c167f42e-b158-462b-966f-d744475140cc", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1189472144-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "38d4c69cad9d40309abf5885ad36939e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "22b8c642-38ad-4c11-9051-145ab3bc54f2", "external-id": "nsx-vlan-transportzone-247", "segmentation_id": 247, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa60eaa2f-cd", "ovs_interfaceid": "a60eaa2f-cdd3-4b07-9a40-d48a4dd95d17", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.983724] env[61987]: DEBUG nova.network.neutron [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Updating instance_info_cache with network_info: [{"id": "114c5cda-5986-410d-a9ef-59f54ab06c10", "address": "fa:16:3e:78:e6:06", "network": {"id": "d5ed1e9e-9033-441e-bfcb-2ac643d7f03f", "bridge": "br-int", "label": "tempest-ServersTestJSON-8466993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e62741abfb4471bbc7e11ae7d407ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39ab9baf-90cd-4fe2-8d56-434f8210fc19", "external-id": "nsx-vlan-transportzone-713", "segmentation_id": 713, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap114c5cda-59", "ovs_interfaceid": "114c5cda-5986-410d-a9ef-59f54ab06c10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.009278] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17c830a2-0e87-4de2-8087-05e10b127222 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.019353] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c398a5-4805-49f1-85d8-5dd95acb55c4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.055608] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a46b5960-d978-4c7c-8365-f228a621cd84 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.064153] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d517e31-725c-46a9-a944-cc936d958c75 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.085424] env[61987]: DEBUG nova.compute.provider_tree [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.194912] env[61987]: DEBUG nova.network.neutron [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Successfully created port: feaf0c72-3782-4054-907b-f98cc72cd02a {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 939.243757] env[61987]: DEBUG oslo_concurrency.lockutils [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "ad9041d9-1e0f-46d6-93f7-efe82197deb6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.244189] env[61987]: DEBUG oslo_concurrency.lockutils [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "ad9041d9-1e0f-46d6-93f7-efe82197deb6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.244439] env[61987]: DEBUG oslo_concurrency.lockutils [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "ad9041d9-1e0f-46d6-93f7-efe82197deb6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.244726] env[61987]: DEBUG oslo_concurrency.lockutils [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "ad9041d9-1e0f-46d6-93f7-efe82197deb6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.244923] env[61987]: DEBUG oslo_concurrency.lockutils [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "ad9041d9-1e0f-46d6-93f7-efe82197deb6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.247650] env[61987]: INFO nova.compute.manager [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Terminating instance [ 939.301920] env[61987]: DEBUG nova.compute.manager [req-490d8628-6c7f-46d9-9765-db6039543df1 req-e8ad1e66-bc79-4844-8e64-e7b939cbb126 service nova] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Received event network-changed-114c5cda-5986-410d-a9ef-59f54ab06c10 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 939.302627] env[61987]: DEBUG nova.compute.manager [req-490d8628-6c7f-46d9-9765-db6039543df1 req-e8ad1e66-bc79-4844-8e64-e7b939cbb126 service nova] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Refreshing instance network info cache due to event network-changed-114c5cda-5986-410d-a9ef-59f54ab06c10. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 939.303242] env[61987]: DEBUG oslo_concurrency.lockutils [req-490d8628-6c7f-46d9-9765-db6039543df1 req-e8ad1e66-bc79-4844-8e64-e7b939cbb126 service nova] Acquiring lock "refresh_cache-f9c82f38-7e06-4949-80a5-6eaa3c01c88c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.305675] env[61987]: INFO nova.compute.manager [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Took 31.95 seconds to build instance. [ 939.317043] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061991, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.379471] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5f9e10f8-d821-4c8d-b8e9-7d9400babbed tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Releasing lock "refresh_cache-5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.488977] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Releasing lock "refresh_cache-f9c82f38-7e06-4949-80a5-6eaa3c01c88c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.488977] env[61987]: DEBUG nova.compute.manager [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Instance network_info: |[{"id": "114c5cda-5986-410d-a9ef-59f54ab06c10", "address": "fa:16:3e:78:e6:06", "network": {"id": "d5ed1e9e-9033-441e-bfcb-2ac643d7f03f", "bridge": "br-int", "label": "tempest-ServersTestJSON-8466993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e62741abfb4471bbc7e11ae7d407ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39ab9baf-90cd-4fe2-8d56-434f8210fc19", "external-id": "nsx-vlan-transportzone-713", "segmentation_id": 713, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap114c5cda-59", "ovs_interfaceid": "114c5cda-5986-410d-a9ef-59f54ab06c10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 939.488977] env[61987]: DEBUG oslo_concurrency.lockutils [req-490d8628-6c7f-46d9-9765-db6039543df1 req-e8ad1e66-bc79-4844-8e64-e7b939cbb126 service nova] Acquired lock "refresh_cache-f9c82f38-7e06-4949-80a5-6eaa3c01c88c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.488977] env[61987]: DEBUG nova.network.neutron [req-490d8628-6c7f-46d9-9765-db6039543df1 req-e8ad1e66-bc79-4844-8e64-e7b939cbb126 service nova] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Refreshing network info cache for port 114c5cda-5986-410d-a9ef-59f54ab06c10 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 939.493821] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:e6:06', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '39ab9baf-90cd-4fe2-8d56-434f8210fc19', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '114c5cda-5986-410d-a9ef-59f54ab06c10', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 939.501328] env[61987]: DEBUG oslo.service.loopingcall [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.502655] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 939.503384] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-69da6e45-ef34-42a8-b258-f0392f99d975 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.523555] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 939.523555] env[61987]: value = "task-1061992" [ 939.523555] env[61987]: _type = "Task" [ 939.523555] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.531797] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061992, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.589603] env[61987]: DEBUG nova.scheduler.client.report [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 939.743875] env[61987]: DEBUG nova.compute.manager [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 939.756028] env[61987]: DEBUG nova.compute.manager [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 939.756028] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 939.756028] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c913a998-1c1f-4ec2-b131-aa87194585ae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.762989] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 939.763559] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-44938942-a327-46b3-8673-244e27a6c26b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.770668] env[61987]: DEBUG oslo_vmware.api [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 939.770668] env[61987]: value = "task-1061993" [ 939.770668] env[61987]: _type = "Task" [ 939.770668] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.782811] env[61987]: DEBUG oslo_vmware.api [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061993, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.800215] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061991, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.57075} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.802327] env[61987]: DEBUG nova.virt.hardware [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 939.802677] env[61987]: DEBUG nova.virt.hardware [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 939.803025] env[61987]: DEBUG nova.virt.hardware [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 939.803392] env[61987]: DEBUG nova.virt.hardware [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 939.803697] env[61987]: DEBUG nova.virt.hardware [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 939.803962] env[61987]: DEBUG nova.virt.hardware [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 939.804292] env[61987]: DEBUG nova.virt.hardware [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 939.804460] env[61987]: DEBUG nova.virt.hardware [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 939.804709] env[61987]: DEBUG nova.virt.hardware [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 939.804934] env[61987]: DEBUG nova.virt.hardware [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 939.805170] env[61987]: DEBUG nova.virt.hardware [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 939.805538] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] a720083b-7b16-468a-a595-bfe8592e8621/a720083b-7b16-468a-a595-bfe8592e8621.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 939.805875] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 939.806641] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e7b79c-b9c7-40d8-bacd-72a453f7f0ab {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.809530] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-696c5adb-4b49-430e-99c8-2f163df8eb05 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.813031] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d225c382-fe17-495e-8416-f049aee00efa tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.463s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.819781] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-191b0478-40ea-4c61-9131-ff8d318b3faf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.825257] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 939.825257] env[61987]: value = "task-1061994" [ 939.825257] env[61987]: _type = "Task" [ 939.825257] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.846614] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061994, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.035232] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1061992, 'name': CreateVM_Task, 'duration_secs': 0.503253} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.037973] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 940.040259] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.040259] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.040259] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 940.040259] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f79f88e3-4889-4b04-8842-79ea5394688f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.044942] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 940.044942] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52b191ab-0a0a-87b7-9d51-70829c54f06f" [ 940.044942] env[61987]: _type = "Task" [ 940.044942] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.053393] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b191ab-0a0a-87b7-9d51-70829c54f06f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.095248] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.025s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.096010] env[61987]: DEBUG nova.compute.manager [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 940.100721] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.833s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.100961] env[61987]: DEBUG nova.objects.instance [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lazy-loading 'resources' on Instance uuid 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.235978] env[61987]: DEBUG nova.network.neutron [req-490d8628-6c7f-46d9-9765-db6039543df1 req-e8ad1e66-bc79-4844-8e64-e7b939cbb126 service nova] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Updated VIF entry in instance network info cache for port 114c5cda-5986-410d-a9ef-59f54ab06c10. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 940.236381] env[61987]: DEBUG nova.network.neutron [req-490d8628-6c7f-46d9-9765-db6039543df1 req-e8ad1e66-bc79-4844-8e64-e7b939cbb126 service nova] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Updating instance_info_cache with network_info: [{"id": "114c5cda-5986-410d-a9ef-59f54ab06c10", "address": "fa:16:3e:78:e6:06", "network": {"id": "d5ed1e9e-9033-441e-bfcb-2ac643d7f03f", "bridge": "br-int", "label": "tempest-ServersTestJSON-8466993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e62741abfb4471bbc7e11ae7d407ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39ab9baf-90cd-4fe2-8d56-434f8210fc19", "external-id": "nsx-vlan-transportzone-713", "segmentation_id": 713, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap114c5cda-59", "ovs_interfaceid": "114c5cda-5986-410d-a9ef-59f54ab06c10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.280932] env[61987]: DEBUG oslo_vmware.api [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061993, 'name': PowerOffVM_Task, 'duration_secs': 0.367463} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.285022] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 940.285022] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 940.285022] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-28c547a0-1048-43dc-86fd-ca1829900431 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.336841] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061994, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.197535} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.337253] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 940.338040] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93c2e752-431d-4a3f-84ab-8d5fb28aa2ba {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.359815] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] a720083b-7b16-468a-a595-bfe8592e8621/a720083b-7b16-468a-a595-bfe8592e8621.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 940.360089] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-14e2a27b-9ad7-4c1e-bb97-48a31a92464b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.379433] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 940.379433] env[61987]: value = "task-1061996" [ 940.379433] env[61987]: _type = "Task" [ 940.379433] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.380844] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 940.381056] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 940.381244] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Deleting the datastore file [datastore2] ad9041d9-1e0f-46d6-93f7-efe82197deb6 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 940.384219] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9d39359c-48de-460a-abe2-14aa31f00cd3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.386185] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f9e10f8-d821-4c8d-b8e9-7d9400babbed tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 940.386417] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9a7f3e20-4f87-4b9e-90d1-f0433d761970 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.392196] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061996, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.394120] env[61987]: DEBUG oslo_vmware.api [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for the task: (returnval){ [ 940.394120] env[61987]: value = "task-1061997" [ 940.394120] env[61987]: _type = "Task" [ 940.394120] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.394363] env[61987]: DEBUG oslo_vmware.api [None req-5f9e10f8-d821-4c8d-b8e9-7d9400babbed tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 940.394363] env[61987]: value = "task-1061998" [ 940.394363] env[61987]: _type = "Task" [ 940.394363] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.405185] env[61987]: DEBUG oslo_vmware.api [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061997, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.407989] env[61987]: DEBUG oslo_vmware.api [None req-5f9e10f8-d821-4c8d-b8e9-7d9400babbed tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061998, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.562560] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b191ab-0a0a-87b7-9d51-70829c54f06f, 'name': SearchDatastore_Task, 'duration_secs': 0.023026} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.562878] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.563133] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 940.563374] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.563518] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.563930] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 940.564033] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3672a4fa-b435-4bc8-b298-77a2bd7f015a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.572281] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 940.572469] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 940.573259] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2648bf52-3ecb-4889-8e73-139588485364 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.578326] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 940.578326] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]529a2617-cf50-d509-c20a-1d1ac95c99fe" [ 940.578326] env[61987]: _type = "Task" [ 940.578326] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.589023] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]529a2617-cf50-d509-c20a-1d1ac95c99fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.604229] env[61987]: DEBUG nova.compute.utils [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 940.610722] env[61987]: DEBUG nova.compute.manager [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 940.610722] env[61987]: DEBUG nova.network.neutron [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 940.705966] env[61987]: DEBUG nova.policy [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0945d6196b8c473da5308e6d784ab479', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b77fbbfc7f6437083f9f30e8fdeeb35', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 940.739017] env[61987]: DEBUG oslo_concurrency.lockutils [req-490d8628-6c7f-46d9-9765-db6039543df1 req-e8ad1e66-bc79-4844-8e64-e7b939cbb126 service nova] Releasing lock "refresh_cache-f9c82f38-7e06-4949-80a5-6eaa3c01c88c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.827212] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a55afd71-232e-4f12-9309-ec4b6cf1bdb7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.835340] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2795d98-6ded-468b-bd1a-f15afd69e749 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.866276] env[61987]: DEBUG nova.network.neutron [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Successfully updated port: feaf0c72-3782-4054-907b-f98cc72cd02a {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 940.867978] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fb65a90-d5a3-406f-ab04-11d8ea037f55 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.875747] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f1a37c-89b1-4589-abe7-4f304c12eab2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.896527] env[61987]: DEBUG nova.compute.provider_tree [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.908409] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061996, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.909428] env[61987]: DEBUG nova.scheduler.client.report [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 940.918718] env[61987]: DEBUG oslo_vmware.api [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Task: {'id': task-1061997, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.266062} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.922026] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 940.922236] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 940.922534] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 940.922679] env[61987]: INFO nova.compute.manager [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Took 1.17 seconds to destroy the instance on the hypervisor. [ 940.922844] env[61987]: DEBUG oslo.service.loopingcall [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.923088] env[61987]: DEBUG oslo_vmware.api [None req-5f9e10f8-d821-4c8d-b8e9-7d9400babbed tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061998, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.923515] env[61987]: DEBUG nova.compute.manager [-] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 940.923615] env[61987]: DEBUG nova.network.neutron [-] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 941.050409] env[61987]: DEBUG nova.network.neutron [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Successfully created port: 9df062f4-8440-4694-9896-000aa92195e2 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 941.089266] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]529a2617-cf50-d509-c20a-1d1ac95c99fe, 'name': SearchDatastore_Task, 'duration_secs': 0.012247} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.090092] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c7b05e3-d8b9-4480-8774-b2fb91fbae06 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.096434] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 941.096434] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5214d1e8-91bc-0fb0-f256-cc668943b759" [ 941.096434] env[61987]: _type = "Task" [ 941.096434] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.104512] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5214d1e8-91bc-0fb0-f256-cc668943b759, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.109369] env[61987]: DEBUG nova.compute.manager [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 941.323942] env[61987]: DEBUG nova.compute.manager [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Received event network-changed-57c4615c-de74-4af1-8b67-1801365b0ea4 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 941.324304] env[61987]: DEBUG nova.compute.manager [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Refreshing instance network info cache due to event network-changed-57c4615c-de74-4af1-8b67-1801365b0ea4. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 941.324640] env[61987]: DEBUG oslo_concurrency.lockutils [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] Acquiring lock "refresh_cache-ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.324841] env[61987]: DEBUG oslo_concurrency.lockutils [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] Acquired lock "refresh_cache-ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.325060] env[61987]: DEBUG nova.network.neutron [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Refreshing network info cache for port 57c4615c-de74-4af1-8b67-1801365b0ea4 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 941.371591] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Acquiring lock "refresh_cache-cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.371753] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Acquired lock "refresh_cache-cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.371912] env[61987]: DEBUG nova.network.neutron [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 941.393213] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061996, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.405402] env[61987]: DEBUG oslo_vmware.api [None req-5f9e10f8-d821-4c8d-b8e9-7d9400babbed tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061998, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.414429] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.314s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.438589] env[61987]: INFO nova.scheduler.client.report [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Deleted allocations for instance 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa [ 941.607929] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5214d1e8-91bc-0fb0-f256-cc668943b759, 'name': SearchDatastore_Task, 'duration_secs': 0.028282} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.608164] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.608321] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] f9c82f38-7e06-4949-80a5-6eaa3c01c88c/f9c82f38-7e06-4949-80a5-6eaa3c01c88c.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 941.608658] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6bf60f81-636c-4948-96d7-639af86d0542 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.621141] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 941.621141] env[61987]: value = "task-1061999" [ 941.621141] env[61987]: _type = "Task" [ 941.621141] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.632957] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061999, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.707620] env[61987]: DEBUG nova.network.neutron [-] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.890838] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1061996, 'name': ReconfigVM_Task, 'duration_secs': 1.322034} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.891147] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Reconfigured VM instance instance-0000005b to attach disk [datastore2] a720083b-7b16-468a-a595-bfe8592e8621/a720083b-7b16-468a-a595-bfe8592e8621.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 941.891544] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d68661a2-1857-4f81-8f17-853f0277f52d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.897840] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 941.897840] env[61987]: value = "task-1062000" [ 941.897840] env[61987]: _type = "Task" [ 941.897840] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.913164] env[61987]: DEBUG oslo_vmware.api [None req-5f9e10f8-d821-4c8d-b8e9-7d9400babbed tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1061998, 'name': PowerOnVM_Task, 'duration_secs': 1.100889} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.913357] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062000, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.914056] env[61987]: DEBUG nova.network.neutron [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 941.915731] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f9e10f8-d821-4c8d-b8e9-7d9400babbed tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 941.915950] env[61987]: DEBUG nova.compute.manager [None req-5f9e10f8-d821-4c8d-b8e9-7d9400babbed tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 941.916734] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f47d761-31a6-4a3a-94dc-f6d8fb1f415a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.948253] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e1b5e4ac-5981-47fe-a505-edc7ad605bb5 tempest-SecurityGroupsTestJSON-725895312 tempest-SecurityGroupsTestJSON-725895312-project-member] Lock "455562c7-9fc1-4625-95b0-5eb8c9c8ccaa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.130s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.123140] env[61987]: DEBUG nova.compute.manager [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 942.134801] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061999, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.152575] env[61987]: DEBUG nova.virt.hardware [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 942.152863] env[61987]: DEBUG nova.virt.hardware [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 942.153080] env[61987]: DEBUG nova.virt.hardware [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 942.153311] env[61987]: DEBUG nova.virt.hardware [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 942.153470] env[61987]: DEBUG nova.virt.hardware [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 942.153626] env[61987]: DEBUG nova.virt.hardware [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 942.153848] env[61987]: DEBUG nova.virt.hardware [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 942.154028] env[61987]: DEBUG nova.virt.hardware [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 942.154209] env[61987]: DEBUG nova.virt.hardware [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 942.154732] env[61987]: DEBUG nova.virt.hardware [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 942.154732] env[61987]: DEBUG nova.virt.hardware [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 942.155542] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f8c816c-ab2d-4002-94c6-cb98c7bee31a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.159111] env[61987]: DEBUG nova.network.neutron [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Updating instance_info_cache with network_info: [{"id": "feaf0c72-3782-4054-907b-f98cc72cd02a", "address": "fa:16:3e:13:f5:64", "network": {"id": "b45b790f-b724-4a9a-b54b-8f8319b13a88", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1186325730-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7f68eb92e124eab9beb7b60ddcc1df3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfeaf0c72-37", "ovs_interfaceid": "feaf0c72-3782-4054-907b-f98cc72cd02a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.165980] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbf49c75-a5d3-4d5f-81b8-1086cba525e0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.187595] env[61987]: DEBUG nova.network.neutron [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Updated VIF entry in instance network info cache for port 57c4615c-de74-4af1-8b67-1801365b0ea4. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 942.188014] env[61987]: DEBUG nova.network.neutron [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Updating instance_info_cache with network_info: [{"id": "57c4615c-de74-4af1-8b67-1801365b0ea4", "address": "fa:16:3e:73:38:cd", "network": {"id": "a9093c43-6bc4-45b1-bdcf-a434759300bd", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-355385599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.210", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7efd6c745f547d3803d5a6184547bda", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ac563aa7-6d7c-4bd5-9241-7b3e11b8f22d", "external-id": "nsx-vlan-transportzone-730", "segmentation_id": 730, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57c4615c-de", "ovs_interfaceid": "57c4615c-de74-4af1-8b67-1801365b0ea4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.212464] env[61987]: INFO nova.compute.manager [-] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Took 1.29 seconds to deallocate network for instance. [ 942.410658] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062000, 'name': Rename_Task, 'duration_secs': 0.147941} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.410947] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 942.411219] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-867c3e1d-e1b0-4cdd-ae89-a4d3fb139131 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.418319] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 942.418319] env[61987]: value = "task-1062001" [ 942.418319] env[61987]: _type = "Task" [ 942.418319] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.430138] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062001, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.635353] env[61987]: DEBUG nova.compute.manager [req-2f7afdee-c605-431e-8a8b-c9249d10388f req-c570bfc2-5755-443f-aca1-f5fe29f77364 service nova] [instance: 865222ae-12da-4e5a-84af-041003b50520] Received event network-vif-plugged-9df062f4-8440-4694-9896-000aa92195e2 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 942.635760] env[61987]: DEBUG oslo_concurrency.lockutils [req-2f7afdee-c605-431e-8a8b-c9249d10388f req-c570bfc2-5755-443f-aca1-f5fe29f77364 service nova] Acquiring lock "865222ae-12da-4e5a-84af-041003b50520-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.636058] env[61987]: DEBUG oslo_concurrency.lockutils [req-2f7afdee-c605-431e-8a8b-c9249d10388f req-c570bfc2-5755-443f-aca1-f5fe29f77364 service nova] Lock "865222ae-12da-4e5a-84af-041003b50520-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.636343] env[61987]: DEBUG oslo_concurrency.lockutils [req-2f7afdee-c605-431e-8a8b-c9249d10388f req-c570bfc2-5755-443f-aca1-f5fe29f77364 service nova] Lock "865222ae-12da-4e5a-84af-041003b50520-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.636639] env[61987]: DEBUG nova.compute.manager [req-2f7afdee-c605-431e-8a8b-c9249d10388f req-c570bfc2-5755-443f-aca1-f5fe29f77364 service nova] [instance: 865222ae-12da-4e5a-84af-041003b50520] No waiting events found dispatching network-vif-plugged-9df062f4-8440-4694-9896-000aa92195e2 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 942.636948] env[61987]: WARNING nova.compute.manager [req-2f7afdee-c605-431e-8a8b-c9249d10388f req-c570bfc2-5755-443f-aca1-f5fe29f77364 service nova] [instance: 865222ae-12da-4e5a-84af-041003b50520] Received unexpected event network-vif-plugged-9df062f4-8440-4694-9896-000aa92195e2 for instance with vm_state building and task_state spawning. [ 942.645561] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1061999, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.733356} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.645916] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] f9c82f38-7e06-4949-80a5-6eaa3c01c88c/f9c82f38-7e06-4949-80a5-6eaa3c01c88c.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 942.646367] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 942.646706] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d4c47228-6dce-46e8-825d-9bd12bb8a6f9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.653091] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 942.653091] env[61987]: value = "task-1062002" [ 942.653091] env[61987]: _type = "Task" [ 942.653091] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.661614] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Releasing lock "refresh_cache-cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.661959] env[61987]: DEBUG nova.compute.manager [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Instance network_info: |[{"id": "feaf0c72-3782-4054-907b-f98cc72cd02a", "address": "fa:16:3e:13:f5:64", "network": {"id": "b45b790f-b724-4a9a-b54b-8f8319b13a88", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1186325730-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7f68eb92e124eab9beb7b60ddcc1df3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfeaf0c72-37", "ovs_interfaceid": "feaf0c72-3782-4054-907b-f98cc72cd02a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 942.662282] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062002, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.662672] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:13:f5:64', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '767a3a48-41d4-4a0c-961d-0024837f63bd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'feaf0c72-3782-4054-907b-f98cc72cd02a', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 942.670653] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Creating folder: Project (e7f68eb92e124eab9beb7b60ddcc1df3). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 942.670917] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3ae1bdf3-328e-4de5-ba29-a71bb3696cf1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.680476] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Created folder: Project (e7f68eb92e124eab9beb7b60ddcc1df3) in parent group-v234219. [ 942.680693] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Creating folder: Instances. Parent ref: group-v234391. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 942.680985] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3fb2e974-1feb-4263-842a-a9e7738cb929 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.689612] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Created folder: Instances in parent group-v234391. [ 942.689824] env[61987]: DEBUG oslo.service.loopingcall [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.690313] env[61987]: DEBUG oslo_concurrency.lockutils [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] Releasing lock "refresh_cache-ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.690549] env[61987]: DEBUG nova.compute.manager [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Received event network-vif-plugged-feaf0c72-3782-4054-907b-f98cc72cd02a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 942.690774] env[61987]: DEBUG oslo_concurrency.lockutils [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] Acquiring lock "cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.691015] env[61987]: DEBUG oslo_concurrency.lockutils [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] Lock "cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.691248] env[61987]: DEBUG oslo_concurrency.lockutils [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] Lock "cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.691446] env[61987]: DEBUG nova.compute.manager [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] No waiting events found dispatching network-vif-plugged-feaf0c72-3782-4054-907b-f98cc72cd02a {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 942.691967] env[61987]: WARNING nova.compute.manager [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Received unexpected event network-vif-plugged-feaf0c72-3782-4054-907b-f98cc72cd02a for instance with vm_state building and task_state spawning. [ 942.691967] env[61987]: DEBUG nova.compute.manager [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Received event network-changed-feaf0c72-3782-4054-907b-f98cc72cd02a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 942.691967] env[61987]: DEBUG nova.compute.manager [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Refreshing instance network info cache due to event network-changed-feaf0c72-3782-4054-907b-f98cc72cd02a. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 942.692233] env[61987]: DEBUG oslo_concurrency.lockutils [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] Acquiring lock "refresh_cache-cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.692391] env[61987]: DEBUG oslo_concurrency.lockutils [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] Acquired lock "refresh_cache-cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.692625] env[61987]: DEBUG nova.network.neutron [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Refreshing network info cache for port feaf0c72-3782-4054-907b-f98cc72cd02a {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 942.693820] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 942.694293] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d7d6f85e-7b3a-47e0-a86b-3fb7db148173 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.719586] env[61987]: DEBUG oslo_concurrency.lockutils [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.719904] env[61987]: DEBUG oslo_concurrency.lockutils [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.720147] env[61987]: DEBUG nova.objects.instance [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lazy-loading 'resources' on Instance uuid ad9041d9-1e0f-46d6-93f7-efe82197deb6 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.721517] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 942.721517] env[61987]: value = "task-1062005" [ 942.721517] env[61987]: _type = "Task" [ 942.721517] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.733773] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062005, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.800585] env[61987]: DEBUG nova.network.neutron [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Successfully updated port: 9df062f4-8440-4694-9896-000aa92195e2 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 942.929628] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062001, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.162821] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062002, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066041} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.163124] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 943.163934] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f19736fc-1e93-47d7-bb20-b4b59dc1eb5f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.187476] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] f9c82f38-7e06-4949-80a5-6eaa3c01c88c/f9c82f38-7e06-4949-80a5-6eaa3c01c88c.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 943.187857] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7c654ef-624f-4aa1-808b-fb2a0a875fb2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.209491] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 943.209491] env[61987]: value = "task-1062006" [ 943.209491] env[61987]: _type = "Task" [ 943.209491] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.219425] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062006, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.233781] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062005, 'name': CreateVM_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.305037] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "refresh_cache-865222ae-12da-4e5a-84af-041003b50520" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.305403] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquired lock "refresh_cache-865222ae-12da-4e5a-84af-041003b50520" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.305553] env[61987]: DEBUG nova.network.neutron [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 943.433449] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062001, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.462651] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d68e63f-114d-4105-ac39-d4d375350019 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.470930] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f03b872-a9e2-4096-bba2-4787c460df6e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.503380] env[61987]: DEBUG nova.network.neutron [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Updated VIF entry in instance network info cache for port feaf0c72-3782-4054-907b-f98cc72cd02a. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 943.503761] env[61987]: DEBUG nova.network.neutron [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Updating instance_info_cache with network_info: [{"id": "feaf0c72-3782-4054-907b-f98cc72cd02a", "address": "fa:16:3e:13:f5:64", "network": {"id": "b45b790f-b724-4a9a-b54b-8f8319b13a88", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1186325730-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7f68eb92e124eab9beb7b60ddcc1df3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfeaf0c72-37", "ovs_interfaceid": "feaf0c72-3782-4054-907b-f98cc72cd02a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.505507] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3fdd439-981d-4f9e-99ec-777171313a46 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.515643] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f547a2-69b3-4d07-9840-161349048c52 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.534460] env[61987]: DEBUG nova.compute.provider_tree [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.719287] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062006, 'name': ReconfigVM_Task, 'duration_secs': 0.357271} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.719619] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Reconfigured VM instance instance-0000005c to attach disk [datastore1] f9c82f38-7e06-4949-80a5-6eaa3c01c88c/f9c82f38-7e06-4949-80a5-6eaa3c01c88c.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 943.720306] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-125e801f-40d4-4fd9-bcd5-603c43cbca48 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.726819] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 943.726819] env[61987]: value = "task-1062007" [ 943.726819] env[61987]: _type = "Task" [ 943.726819] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.737168] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062005, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.740910] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062007, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.858179] env[61987]: DEBUG nova.network.neutron [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 943.932929] env[61987]: DEBUG oslo_vmware.api [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062001, 'name': PowerOnVM_Task, 'duration_secs': 1.219886} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.933288] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 943.933602] env[61987]: INFO nova.compute.manager [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Took 11.44 seconds to spawn the instance on the hypervisor. [ 943.933602] env[61987]: DEBUG nova.compute.manager [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 943.937482] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9571e0d-ba29-4e02-9321-1905f74a6a9e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.008917] env[61987]: DEBUG oslo_concurrency.lockutils [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] Releasing lock "refresh_cache-cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.009208] env[61987]: DEBUG nova.compute.manager [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Received event network-vif-deleted-d857e05f-901f-4f86-9f9e-33aed6d5fec9 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 944.009397] env[61987]: INFO nova.compute.manager [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Neutron deleted interface d857e05f-901f-4f86-9f9e-33aed6d5fec9; detaching it from the instance and deleting it from the info cache [ 944.009577] env[61987]: DEBUG nova.network.neutron [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.037353] env[61987]: DEBUG nova.scheduler.client.report [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 944.164028] env[61987]: DEBUG nova.network.neutron [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Updating instance_info_cache with network_info: [{"id": "9df062f4-8440-4694-9896-000aa92195e2", "address": "fa:16:3e:15:06:44", "network": {"id": "26d453d1-8edf-45e1-9e49-2a3225de6633", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1400564876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b77fbbfc7f6437083f9f30e8fdeeb35", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9df062f4-84", "ovs_interfaceid": "9df062f4-8440-4694-9896-000aa92195e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.241823] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062007, 'name': Rename_Task, 'duration_secs': 0.149305} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.242256] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 944.242833] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062005, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.243056] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fa675ecf-086d-437e-8ea5-ba39e054e97f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.249721] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 944.249721] env[61987]: value = "task-1062008" [ 944.249721] env[61987]: _type = "Task" [ 944.249721] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.258936] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062008, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.459385] env[61987]: INFO nova.compute.manager [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Took 33.55 seconds to build instance. [ 944.513092] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ac46b759-19fd-4ddc-9ddb-5050f794ddc9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.522483] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c025c2-cc70-44f5-b834-50c1629cd667 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.543026] env[61987]: DEBUG oslo_concurrency.lockutils [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.823s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.556511] env[61987]: DEBUG nova.compute.manager [req-b7d8b921-d4b6-4c98-8b07-a37f5a5ca8d8 req-a88b936c-33f0-4be6-a1a7-69020e730961 service nova] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Detach interface failed, port_id=d857e05f-901f-4f86-9f9e-33aed6d5fec9, reason: Instance ad9041d9-1e0f-46d6-93f7-efe82197deb6 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 944.575387] env[61987]: INFO nova.scheduler.client.report [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Deleted allocations for instance ad9041d9-1e0f-46d6-93f7-efe82197deb6 [ 944.666527] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Releasing lock "refresh_cache-865222ae-12da-4e5a-84af-041003b50520" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.666855] env[61987]: DEBUG nova.compute.manager [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Instance network_info: |[{"id": "9df062f4-8440-4694-9896-000aa92195e2", "address": "fa:16:3e:15:06:44", "network": {"id": "26d453d1-8edf-45e1-9e49-2a3225de6633", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1400564876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b77fbbfc7f6437083f9f30e8fdeeb35", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9df062f4-84", "ovs_interfaceid": "9df062f4-8440-4694-9896-000aa92195e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 944.667576] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:15:06:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3734b156-0f7d-4721-b23c-d000412ec2eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9df062f4-8440-4694-9896-000aa92195e2', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 944.678371] env[61987]: DEBUG oslo.service.loopingcall [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 944.679957] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 865222ae-12da-4e5a-84af-041003b50520] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 944.681549] env[61987]: DEBUG nova.compute.manager [req-9f72fe2a-57b8-4b59-9f26-5ec3c67a28f9 req-e4f71ee3-533e-460c-b14c-321d5a113978 service nova] [instance: 865222ae-12da-4e5a-84af-041003b50520] Received event network-changed-9df062f4-8440-4694-9896-000aa92195e2 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 944.681719] env[61987]: DEBUG nova.compute.manager [req-9f72fe2a-57b8-4b59-9f26-5ec3c67a28f9 req-e4f71ee3-533e-460c-b14c-321d5a113978 service nova] [instance: 865222ae-12da-4e5a-84af-041003b50520] Refreshing instance network info cache due to event network-changed-9df062f4-8440-4694-9896-000aa92195e2. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 944.682444] env[61987]: DEBUG oslo_concurrency.lockutils [req-9f72fe2a-57b8-4b59-9f26-5ec3c67a28f9 req-e4f71ee3-533e-460c-b14c-321d5a113978 service nova] Acquiring lock "refresh_cache-865222ae-12da-4e5a-84af-041003b50520" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.682565] env[61987]: DEBUG oslo_concurrency.lockutils [req-9f72fe2a-57b8-4b59-9f26-5ec3c67a28f9 req-e4f71ee3-533e-460c-b14c-321d5a113978 service nova] Acquired lock "refresh_cache-865222ae-12da-4e5a-84af-041003b50520" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.682861] env[61987]: DEBUG nova.network.neutron [req-9f72fe2a-57b8-4b59-9f26-5ec3c67a28f9 req-e4f71ee3-533e-460c-b14c-321d5a113978 service nova] [instance: 865222ae-12da-4e5a-84af-041003b50520] Refreshing network info cache for port 9df062f4-8440-4694-9896-000aa92195e2 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 944.683862] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-06d5d5d0-7c8d-47f9-9a8b-a2000f7b53d2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.706853] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 944.706853] env[61987]: value = "task-1062009" [ 944.706853] env[61987]: _type = "Task" [ 944.706853] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.715511] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062009, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.738406] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062005, 'name': CreateVM_Task, 'duration_secs': 1.700633} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.738593] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 944.739355] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.739468] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.739796] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 944.740068] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f41f2c5a-912e-4beb-9c28-e4f92d33d83b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.744739] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Waiting for the task: (returnval){ [ 944.744739] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52b50ea3-9dd1-2731-25a5-6f9012d12a1e" [ 944.744739] env[61987]: _type = "Task" [ 944.744739] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.753154] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b50ea3-9dd1-2731-25a5-6f9012d12a1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.761406] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062008, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.961692] env[61987]: DEBUG oslo_concurrency.lockutils [None req-883164d7-949a-48fa-be96-0da86d442edf tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "a720083b-7b16-468a-a595-bfe8592e8621" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.057s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.087835] env[61987]: DEBUG oslo_concurrency.lockutils [None req-52cb72f2-304f-4d16-a517-178f86dd023b tempest-ServerDiskConfigTestJSON-1955528946 tempest-ServerDiskConfigTestJSON-1955528946-project-member] Lock "ad9041d9-1e0f-46d6-93f7-efe82197deb6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.844s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.147326] env[61987]: DEBUG oslo_concurrency.lockutils [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "876b74ae-cae8-44a9-96ba-e2869ed43be6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.147603] env[61987]: DEBUG oslo_concurrency.lockutils [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "876b74ae-cae8-44a9-96ba-e2869ed43be6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.147819] env[61987]: DEBUG oslo_concurrency.lockutils [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "876b74ae-cae8-44a9-96ba-e2869ed43be6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.148039] env[61987]: DEBUG oslo_concurrency.lockutils [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "876b74ae-cae8-44a9-96ba-e2869ed43be6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.148203] env[61987]: DEBUG oslo_concurrency.lockutils [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "876b74ae-cae8-44a9-96ba-e2869ed43be6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.150500] env[61987]: INFO nova.compute.manager [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Terminating instance [ 945.217522] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062009, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.254458] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b50ea3-9dd1-2731-25a5-6f9012d12a1e, 'name': SearchDatastore_Task, 'duration_secs': 0.010753} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.257547] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.257755] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 945.257995] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.258166] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.258347] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 945.260601] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-375daa6b-a48b-4950-b093-a50043e49375 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.267226] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062008, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.273069] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 945.273265] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 945.273970] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50429a99-02d0-41fd-a8e1-771ae822a686 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.278981] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Waiting for the task: (returnval){ [ 945.278981] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52fbcb61-9757-b237-8801-91701ad330a4" [ 945.278981] env[61987]: _type = "Task" [ 945.278981] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.286315] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52fbcb61-9757-b237-8801-91701ad330a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.654662] env[61987]: DEBUG nova.compute.manager [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 945.654917] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 945.656037] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9d0c326-fdb7-4588-93e2-11f613a60834 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.663731] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 945.663979] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-24548f74-db24-41d0-b7e7-a44b5af5574f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.670204] env[61987]: DEBUG oslo_vmware.api [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 945.670204] env[61987]: value = "task-1062010" [ 945.670204] env[61987]: _type = "Task" [ 945.670204] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.678736] env[61987]: DEBUG oslo_vmware.api [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1062010, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.683612] env[61987]: DEBUG nova.network.neutron [req-9f72fe2a-57b8-4b59-9f26-5ec3c67a28f9 req-e4f71ee3-533e-460c-b14c-321d5a113978 service nova] [instance: 865222ae-12da-4e5a-84af-041003b50520] Updated VIF entry in instance network info cache for port 9df062f4-8440-4694-9896-000aa92195e2. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 945.684050] env[61987]: DEBUG nova.network.neutron [req-9f72fe2a-57b8-4b59-9f26-5ec3c67a28f9 req-e4f71ee3-533e-460c-b14c-321d5a113978 service nova] [instance: 865222ae-12da-4e5a-84af-041003b50520] Updating instance_info_cache with network_info: [{"id": "9df062f4-8440-4694-9896-000aa92195e2", "address": "fa:16:3e:15:06:44", "network": {"id": "26d453d1-8edf-45e1-9e49-2a3225de6633", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1400564876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b77fbbfc7f6437083f9f30e8fdeeb35", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9df062f4-84", "ovs_interfaceid": "9df062f4-8440-4694-9896-000aa92195e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.717547] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062009, 'name': CreateVM_Task, 'duration_secs': 0.562228} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.717718] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 865222ae-12da-4e5a-84af-041003b50520] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 945.718439] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.718617] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.718944] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 945.719237] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3edcbbfa-dd9a-43e0-8247-313cd885ccd1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.725426] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 945.725426] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5287de8d-901a-f402-3de7-43fa8c9b324e" [ 945.725426] env[61987]: _type = "Task" [ 945.725426] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.733916] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5287de8d-901a-f402-3de7-43fa8c9b324e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.766184] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062008, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.793416] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52fbcb61-9757-b237-8801-91701ad330a4, 'name': SearchDatastore_Task, 'duration_secs': 0.020486} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.793945] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69f0f759-d6e3-4f68-bb6d-aef31178d851 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.802881] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Waiting for the task: (returnval){ [ 945.802881] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5218a315-a755-395d-71c9-665bd04aec24" [ 945.802881] env[61987]: _type = "Task" [ 945.802881] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.817515] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5218a315-a755-395d-71c9-665bd04aec24, 'name': SearchDatastore_Task, 'duration_secs': 0.010481} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.818043] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.818699] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0/cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 945.819179] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a7ec0c64-f9b4-4328-ab9d-edcdf60622b7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.829234] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Waiting for the task: (returnval){ [ 945.829234] env[61987]: value = "task-1062011" [ 945.829234] env[61987]: _type = "Task" [ 945.829234] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.837333] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062011, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.983267] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1a7a6aa-33ff-43ca-8a1d-d2baf6eeee85 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.990398] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1cc4a606-7b1d-4645-baf9-25eab87d1c4d tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Suspending the VM {{(pid=61987) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 945.990650] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-4ac56a22-7921-4420-8aff-209a3c7cf82c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.997173] env[61987]: DEBUG oslo_vmware.api [None req-1cc4a606-7b1d-4645-baf9-25eab87d1c4d tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 945.997173] env[61987]: value = "task-1062012" [ 945.997173] env[61987]: _type = "Task" [ 945.997173] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.007692] env[61987]: DEBUG oslo_vmware.api [None req-1cc4a606-7b1d-4645-baf9-25eab87d1c4d tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062012, 'name': SuspendVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.183971] env[61987]: DEBUG oslo_vmware.api [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1062010, 'name': PowerOffVM_Task, 'duration_secs': 0.177817} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.185775] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 946.186012] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 946.186645] env[61987]: DEBUG oslo_concurrency.lockutils [req-9f72fe2a-57b8-4b59-9f26-5ec3c67a28f9 req-e4f71ee3-533e-460c-b14c-321d5a113978 service nova] Releasing lock "refresh_cache-865222ae-12da-4e5a-84af-041003b50520" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.186969] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b8880072-73af-4e83-a237-d0d9424d4590 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.237668] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5287de8d-901a-f402-3de7-43fa8c9b324e, 'name': SearchDatastore_Task, 'duration_secs': 0.01148} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.238242] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.238487] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 946.238782] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.238961] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.239421] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 946.239760] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fb6b8f56-f13c-498b-bac3-441f92557516 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.255620] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 946.255868] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 946.261191] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b7385cd-8fc5-4394-b470-f2b61cd40c59 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.264256] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 946.264535] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 946.264766] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Deleting the datastore file [datastore2] 876b74ae-cae8-44a9-96ba-e2869ed43be6 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 946.264980] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8fc62068-26fc-42c4-8b9d-18ba4e7bb9fc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.270732] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 946.270732] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]527fc789-d1e4-39cb-53b8-79b1b90cb812" [ 946.270732] env[61987]: _type = "Task" [ 946.270732] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.274334] env[61987]: DEBUG oslo_vmware.api [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062008, 'name': PowerOnVM_Task, 'duration_secs': 1.654556} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.279315] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 946.279571] env[61987]: INFO nova.compute.manager [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Took 9.61 seconds to spawn the instance on the hypervisor. [ 946.279760] env[61987]: DEBUG nova.compute.manager [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 946.280206] env[61987]: DEBUG oslo_vmware.api [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 946.280206] env[61987]: value = "task-1062014" [ 946.280206] env[61987]: _type = "Task" [ 946.280206] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.281051] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-815e76a6-122a-4090-bade-75ef3d779193 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.293460] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]527fc789-d1e4-39cb-53b8-79b1b90cb812, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.304284] env[61987]: DEBUG oslo_vmware.api [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1062014, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.338359] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062011, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.508730] env[61987]: DEBUG oslo_vmware.api [None req-1cc4a606-7b1d-4645-baf9-25eab87d1c4d tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062012, 'name': SuspendVM_Task} progress is 50%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.785653] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]527fc789-d1e4-39cb-53b8-79b1b90cb812, 'name': SearchDatastore_Task, 'duration_secs': 0.053274} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.790959] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1151c021-56c0-4440-8675-eb5c5c8606aa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.804888] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 946.804888] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52145585-b49d-119b-ab80-f58321dbd09c" [ 946.804888] env[61987]: _type = "Task" [ 946.804888] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.815191] env[61987]: DEBUG oslo_vmware.api [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1062014, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.266971} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.815191] env[61987]: INFO nova.compute.manager [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Took 21.05 seconds to build instance. [ 946.815191] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 946.815191] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 946.815191] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 946.815191] env[61987]: INFO nova.compute.manager [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Took 1.16 seconds to destroy the instance on the hypervisor. [ 946.815191] env[61987]: DEBUG oslo.service.loopingcall [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.822615] env[61987]: DEBUG nova.compute.manager [-] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 946.823089] env[61987]: DEBUG nova.network.neutron [-] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 946.835024] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52145585-b49d-119b-ab80-f58321dbd09c, 'name': SearchDatastore_Task, 'duration_secs': 0.031761} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.839037] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.839417] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 865222ae-12da-4e5a-84af-041003b50520/865222ae-12da-4e5a-84af-041003b50520.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 946.839979] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5e45c32c-0757-4c59-976e-49374c67c5f7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.846675] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062011, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534251} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.848086] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0/cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 946.848315] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 946.848623] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 946.848623] env[61987]: value = "task-1062015" [ 946.848623] env[61987]: _type = "Task" [ 946.848623] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.848813] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cd8de243-5631-4ab5-b53c-0c26218c4509 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.858168] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062015, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.859473] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Waiting for the task: (returnval){ [ 946.859473] env[61987]: value = "task-1062016" [ 946.859473] env[61987]: _type = "Task" [ 946.859473] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.866970] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062016, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.008595] env[61987]: DEBUG oslo_vmware.api [None req-1cc4a606-7b1d-4645-baf9-25eab87d1c4d tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062012, 'name': SuspendVM_Task, 'duration_secs': 0.714403} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.008945] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1cc4a606-7b1d-4645-baf9-25eab87d1c4d tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Suspended the VM {{(pid=61987) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 947.009100] env[61987]: DEBUG nova.compute.manager [None req-1cc4a606-7b1d-4645-baf9-25eab87d1c4d tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 947.009902] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87934b74-178f-45f2-877c-b1bb728afcc9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.317792] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c7789159-c7c6-4a66-a404-1d3d2eb5b32b tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "f9c82f38-7e06-4949-80a5-6eaa3c01c88c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.562s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.346282] env[61987]: DEBUG nova.compute.manager [req-02dda8b8-fedf-44bb-a549-377bd59b73bb req-b8a1f488-1617-4d56-b85c-ea0df681b906 service nova] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Received event network-vif-deleted-acb22031-7e46-4139-8097-e06110e2fb54 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 947.346828] env[61987]: INFO nova.compute.manager [req-02dda8b8-fedf-44bb-a549-377bd59b73bb req-b8a1f488-1617-4d56-b85c-ea0df681b906 service nova] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Neutron deleted interface acb22031-7e46-4139-8097-e06110e2fb54; detaching it from the instance and deleting it from the info cache [ 947.346828] env[61987]: DEBUG nova.network.neutron [req-02dda8b8-fedf-44bb-a549-377bd59b73bb req-b8a1f488-1617-4d56-b85c-ea0df681b906 service nova] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.362316] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062015, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.372171] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062016, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071397} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.372500] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 947.373261] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c588787d-ae6a-4849-bed8-ce8ba5021ae5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.407470] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0/cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 947.407470] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c1f83a4-c06a-438e-958f-ee47261b8eaf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.435025] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Waiting for the task: (returnval){ [ 947.435025] env[61987]: value = "task-1062017" [ 947.435025] env[61987]: _type = "Task" [ 947.435025] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.447115] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062017, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.814483] env[61987]: DEBUG nova.network.neutron [-] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.849853] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-da42c1b6-20dc-4202-b690-037ddf16f959 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.865721] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062015, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.870676] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9dcd56c-5242-491d-b04e-441eda0f54cd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.909602] env[61987]: DEBUG nova.compute.manager [req-02dda8b8-fedf-44bb-a549-377bd59b73bb req-b8a1f488-1617-4d56-b85c-ea0df681b906 service nova] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Detach interface failed, port_id=acb22031-7e46-4139-8097-e06110e2fb54, reason: Instance 876b74ae-cae8-44a9-96ba-e2869ed43be6 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 947.950987] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062017, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.318403] env[61987]: INFO nova.compute.manager [-] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Took 1.50 seconds to deallocate network for instance. [ 948.369692] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062015, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.143724} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.371020] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 865222ae-12da-4e5a-84af-041003b50520/865222ae-12da-4e5a-84af-041003b50520.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 948.371225] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 948.374343] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8a7789c8-983b-4f06-8d9f-81107769f1de {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.381565] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 948.381565] env[61987]: value = "task-1062018" [ 948.381565] env[61987]: _type = "Task" [ 948.381565] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.391878] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062018, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.448243] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062017, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.487638] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "f9c82f38-7e06-4949-80a5-6eaa3c01c88c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.488221] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "f9c82f38-7e06-4949-80a5-6eaa3c01c88c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.488635] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "f9c82f38-7e06-4949-80a5-6eaa3c01c88c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.490842] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "f9c82f38-7e06-4949-80a5-6eaa3c01c88c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.490842] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "f9c82f38-7e06-4949-80a5-6eaa3c01c88c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.496125] env[61987]: INFO nova.compute.manager [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Terminating instance [ 948.788304] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "a720083b-7b16-468a-a595-bfe8592e8621" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.789098] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "a720083b-7b16-468a-a595-bfe8592e8621" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.789098] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "a720083b-7b16-468a-a595-bfe8592e8621-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.789098] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "a720083b-7b16-468a-a595-bfe8592e8621-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.789247] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "a720083b-7b16-468a-a595-bfe8592e8621-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.792115] env[61987]: INFO nova.compute.manager [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Terminating instance [ 948.828607] env[61987]: DEBUG oslo_concurrency.lockutils [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.828607] env[61987]: DEBUG oslo_concurrency.lockutils [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.828607] env[61987]: DEBUG nova.objects.instance [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lazy-loading 'resources' on Instance uuid 876b74ae-cae8-44a9-96ba-e2869ed43be6 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.879555] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Acquiring lock "d11104e7-4a5c-44c3-bd92-d36c587da794" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.879915] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Lock "d11104e7-4a5c-44c3-bd92-d36c587da794" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.880220] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Acquiring lock "d11104e7-4a5c-44c3-bd92-d36c587da794-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.880469] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Lock "d11104e7-4a5c-44c3-bd92-d36c587da794-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.881528] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Lock "d11104e7-4a5c-44c3-bd92-d36c587da794-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.886856] env[61987]: INFO nova.compute.manager [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Terminating instance [ 948.895383] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062018, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070175} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.896311] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 948.897108] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfbe91d5-891f-4a2e-a681-fd9bf8b1d95d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.922743] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] 865222ae-12da-4e5a-84af-041003b50520/865222ae-12da-4e5a-84af-041003b50520.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 948.923164] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32f40a2f-a8dc-4edb-9a58-065b0ea1b1cd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.947413] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062017, 'name': ReconfigVM_Task, 'duration_secs': 1.021175} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.948709] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Reconfigured VM instance instance-0000005d to attach disk [datastore1] cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0/cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 948.949439] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 948.949439] env[61987]: value = "task-1062019" [ 948.949439] env[61987]: _type = "Task" [ 948.949439] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.949641] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ccfa0c2c-cd73-4c5c-b1b1-5cd34917b6b2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.958878] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062019, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.960235] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Waiting for the task: (returnval){ [ 948.960235] env[61987]: value = "task-1062020" [ 948.960235] env[61987]: _type = "Task" [ 948.960235] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.967666] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062020, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.004160] env[61987]: DEBUG nova.compute.manager [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 949.004423] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 949.005413] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe12cceb-9cb5-4cff-b28d-66cc5fc70933 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.013052] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 949.013316] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1dcb1cc6-e2b5-473d-a4ba-5afb14eba591 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.018503] env[61987]: DEBUG oslo_vmware.api [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 949.018503] env[61987]: value = "task-1062021" [ 949.018503] env[61987]: _type = "Task" [ 949.018503] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.027157] env[61987]: DEBUG oslo_vmware.api [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062021, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.295708] env[61987]: DEBUG nova.compute.manager [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 949.296025] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 949.298902] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5742d21-1995-4dcc-b622-15ff5ec071cd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.307349] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 949.307408] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b1a440f6-6884-4e4f-85c3-75a7c40adf3b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.346019] env[61987]: DEBUG oslo_concurrency.lockutils [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Acquiring lock "70702eab-b510-4946-a426-92902b591481" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.346019] env[61987]: DEBUG oslo_concurrency.lockutils [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Lock "70702eab-b510-4946-a426-92902b591481" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.372205] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 949.372452] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 949.372646] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Deleting the datastore file [datastore2] a720083b-7b16-468a-a595-bfe8592e8621 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 949.376039] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-824d6f6e-60fb-47bc-8f58-fdf2733112d2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.385156] env[61987]: DEBUG oslo_vmware.api [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 949.385156] env[61987]: value = "task-1062023" [ 949.385156] env[61987]: _type = "Task" [ 949.385156] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.393188] env[61987]: DEBUG nova.compute.manager [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 949.393443] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 949.398776] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77a714c6-d926-41e1-86f6-ffb76adca16e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.402451] env[61987]: DEBUG oslo_vmware.api [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062023, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.409325] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 949.409587] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e8a049cf-0fbd-4bbf-967c-071bdb824e62 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.415293] env[61987]: DEBUG oslo_vmware.api [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Waiting for the task: (returnval){ [ 949.415293] env[61987]: value = "task-1062024" [ 949.415293] env[61987]: _type = "Task" [ 949.415293] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.429053] env[61987]: DEBUG oslo_vmware.api [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Task: {'id': task-1062024, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.462097] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062019, 'name': ReconfigVM_Task, 'duration_secs': 0.494311} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.467905] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Reconfigured VM instance instance-0000005e to attach disk [datastore1] 865222ae-12da-4e5a-84af-041003b50520/865222ae-12da-4e5a-84af-041003b50520.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 949.469248] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3a7556e0-c7dd-432f-8999-2d73a2b24c04 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.476354] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062020, 'name': Rename_Task, 'duration_secs': 0.247034} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.477769] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 949.478107] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 949.478107] env[61987]: value = "task-1062025" [ 949.478107] env[61987]: _type = "Task" [ 949.478107] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.478317] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-928a075f-1fd9-441b-a3fc-12c9cb31951b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.491954] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062025, 'name': Rename_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.493975] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Waiting for the task: (returnval){ [ 949.493975] env[61987]: value = "task-1062026" [ 949.493975] env[61987]: _type = "Task" [ 949.493975] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.506012] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062026, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.528507] env[61987]: DEBUG oslo_vmware.api [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062021, 'name': PowerOffVM_Task, 'duration_secs': 0.271474} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.528853] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 949.529057] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 949.532084] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-107c9ce6-fd39-463e-a9bb-080bf76167ff {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.577292] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f2818f7-9d3a-4804-9d88-c5c2d2647c2e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.587499] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb8073b3-3381-4ebf-8b7e-b6239f2559e1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.631065] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eaa9cba-1adf-493f-b2cd-6ccccc3cc777 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.631407] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 949.631718] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 949.631978] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Deleting the datastore file [datastore1] f9c82f38-7e06-4949-80a5-6eaa3c01c88c {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 949.632265] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-47e2555b-3a64-4e1f-a838-5a5b091a318d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.642410] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f75063ea-b91c-4878-9ff2-c6fd38ea218b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.647117] env[61987]: DEBUG oslo_vmware.api [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 949.647117] env[61987]: value = "task-1062028" [ 949.647117] env[61987]: _type = "Task" [ 949.647117] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.660617] env[61987]: DEBUG nova.compute.provider_tree [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 949.667785] env[61987]: DEBUG oslo_vmware.api [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062028, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.848050] env[61987]: DEBUG nova.compute.manager [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 949.895828] env[61987]: DEBUG oslo_vmware.api [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062023, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.272289} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.896195] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 949.896381] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 949.896517] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 949.897123] env[61987]: INFO nova.compute.manager [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Took 0.60 seconds to destroy the instance on the hypervisor. [ 949.897123] env[61987]: DEBUG oslo.service.loopingcall [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 949.897123] env[61987]: DEBUG nova.compute.manager [-] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 949.897323] env[61987]: DEBUG nova.network.neutron [-] [instance: a720083b-7b16-468a-a595-bfe8592e8621] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 949.925503] env[61987]: DEBUG oslo_vmware.api [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Task: {'id': task-1062024, 'name': PowerOffVM_Task, 'duration_secs': 0.230095} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.925821] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 949.926035] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 949.926307] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e11666e0-7240-4bdf-b8a5-418b2df0eee5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.987809] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 949.988066] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 949.988305] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Deleting the datastore file [datastore2] d11104e7-4a5c-44c3-bd92-d36c587da794 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 949.992022] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e972d208-8f1c-4c19-babe-4ae8ea1eca69 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.994949] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062025, 'name': Rename_Task, 'duration_secs': 0.212027} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.001431] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 950.001431] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eddc2208-a15c-49ca-8791-216fece1c255 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.001655] env[61987]: DEBUG oslo_vmware.api [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Waiting for the task: (returnval){ [ 950.001655] env[61987]: value = "task-1062030" [ 950.001655] env[61987]: _type = "Task" [ 950.001655] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.011034] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062026, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.011389] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 950.011389] env[61987]: value = "task-1062031" [ 950.011389] env[61987]: _type = "Task" [ 950.011389] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.018604] env[61987]: DEBUG oslo_vmware.api [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Task: {'id': task-1062030, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.024695] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062031, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.160437] env[61987]: DEBUG oslo_vmware.api [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062028, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.207186} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.161539] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.161539] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 950.161539] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 950.161539] env[61987]: INFO nova.compute.manager [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Took 1.16 seconds to destroy the instance on the hypervisor. [ 950.161763] env[61987]: DEBUG oslo.service.loopingcall [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.161815] env[61987]: DEBUG nova.compute.manager [-] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 950.162067] env[61987]: DEBUG nova.network.neutron [-] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 950.164539] env[61987]: DEBUG nova.scheduler.client.report [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 950.298862] env[61987]: DEBUG nova.compute.manager [req-8bf13446-fc2e-44a6-8999-66676c85fd4d req-227eda67-188d-4158-b88b-1a6ad693e789 service nova] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Received event network-vif-deleted-f0bccd00-ae35-45ee-af83-b29875006ee1 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 950.299210] env[61987]: INFO nova.compute.manager [req-8bf13446-fc2e-44a6-8999-66676c85fd4d req-227eda67-188d-4158-b88b-1a6ad693e789 service nova] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Neutron deleted interface f0bccd00-ae35-45ee-af83-b29875006ee1; detaching it from the instance and deleting it from the info cache [ 950.299414] env[61987]: DEBUG nova.network.neutron [req-8bf13446-fc2e-44a6-8999-66676c85fd4d req-227eda67-188d-4158-b88b-1a6ad693e789 service nova] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.381130] env[61987]: DEBUG oslo_concurrency.lockutils [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.483574] env[61987]: DEBUG nova.compute.manager [req-a81dc995-2cc3-4e73-ba00-db3278dffdb5 req-887d0d64-687c-4409-9067-3d7d8965d908 service nova] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Received event network-vif-deleted-114c5cda-5986-410d-a9ef-59f54ab06c10 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 950.483809] env[61987]: INFO nova.compute.manager [req-a81dc995-2cc3-4e73-ba00-db3278dffdb5 req-887d0d64-687c-4409-9067-3d7d8965d908 service nova] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Neutron deleted interface 114c5cda-5986-410d-a9ef-59f54ab06c10; detaching it from the instance and deleting it from the info cache [ 950.483993] env[61987]: DEBUG nova.network.neutron [req-a81dc995-2cc3-4e73-ba00-db3278dffdb5 req-887d0d64-687c-4409-9067-3d7d8965d908 service nova] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.505291] env[61987]: DEBUG oslo_vmware.api [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062026, 'name': PowerOnVM_Task, 'duration_secs': 0.604624} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.508845] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 950.509137] env[61987]: INFO nova.compute.manager [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Took 10.76 seconds to spawn the instance on the hypervisor. [ 950.509396] env[61987]: DEBUG nova.compute.manager [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 950.510561] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d728fd-5a89-45c5-9fc7-a126b59a8151 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.527038] env[61987]: DEBUG oslo_vmware.api [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Task: {'id': task-1062030, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.224873} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.529441] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.529670] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 950.529942] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 950.530179] env[61987]: INFO nova.compute.manager [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Took 1.14 seconds to destroy the instance on the hypervisor. [ 950.530537] env[61987]: DEBUG oslo.service.loopingcall [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.536948] env[61987]: DEBUG nova.compute.manager [-] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 950.537099] env[61987]: DEBUG nova.network.neutron [-] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 950.538857] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062031, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.672232] env[61987]: DEBUG oslo_concurrency.lockutils [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.845s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.674688] env[61987]: DEBUG oslo_concurrency.lockutils [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.294s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.676291] env[61987]: INFO nova.compute.claims [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 950.696716] env[61987]: INFO nova.scheduler.client.report [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Deleted allocations for instance 876b74ae-cae8-44a9-96ba-e2869ed43be6 [ 950.764152] env[61987]: DEBUG nova.network.neutron [-] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.802448] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-efce8a0d-af43-4d05-aada-62abdf411956 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.811713] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faec9ed7-344c-4baf-85ac-388b6261f746 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.841695] env[61987]: DEBUG nova.compute.manager [req-8bf13446-fc2e-44a6-8999-66676c85fd4d req-227eda67-188d-4158-b88b-1a6ad693e789 service nova] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Detach interface failed, port_id=f0bccd00-ae35-45ee-af83-b29875006ee1, reason: Instance a720083b-7b16-468a-a595-bfe8592e8621 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 950.924473] env[61987]: DEBUG nova.network.neutron [-] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.987416] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ef7c1edf-31ad-4812-a518-4954d6102dfa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.001903] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab3e3690-6411-49d8-9a9c-ab0b39368cc1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.025946] env[61987]: DEBUG oslo_vmware.api [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062031, 'name': PowerOnVM_Task, 'duration_secs': 0.541762} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.035507] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 951.035836] env[61987]: INFO nova.compute.manager [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Took 8.91 seconds to spawn the instance on the hypervisor. [ 951.036185] env[61987]: DEBUG nova.compute.manager [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 951.036513] env[61987]: DEBUG nova.compute.manager [req-a81dc995-2cc3-4e73-ba00-db3278dffdb5 req-887d0d64-687c-4409-9067-3d7d8965d908 service nova] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Detach interface failed, port_id=114c5cda-5986-410d-a9ef-59f54ab06c10, reason: Instance f9c82f38-7e06-4949-80a5-6eaa3c01c88c could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 951.037385] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd7e3f8-078e-47f9-b9df-923c03d4aa49 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.058601] env[61987]: INFO nova.compute.manager [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Took 22.57 seconds to build instance. [ 951.203575] env[61987]: DEBUG oslo_concurrency.lockutils [None req-265bf980-d8f2-483b-b64c-8346b9338dea tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "876b74ae-cae8-44a9-96ba-e2869ed43be6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.056s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.267154] env[61987]: INFO nova.compute.manager [-] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Took 1.37 seconds to deallocate network for instance. [ 951.427882] env[61987]: INFO nova.compute.manager [-] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Took 1.27 seconds to deallocate network for instance. [ 951.562408] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8842c66-344f-4652-89fb-e37d00e6a97b tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Lock "cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.091s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.568655] env[61987]: INFO nova.compute.manager [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Took 16.03 seconds to build instance. [ 951.775178] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.929421] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48edc967-198a-4fe2-b713-fa75ab1d4e61 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.935014] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.944201] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8b61448-cd61-4bc3-8069-992712d4e481 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.979912] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c8755f-12b4-4095-a5e9-24681f2def96 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.987850] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc77c6a6-23ac-416d-b54d-d92b00d6326d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.992142] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "286f24dd-2239-434e-8ae7-15164a0ea8b0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.992402] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "286f24dd-2239-434e-8ae7-15164a0ea8b0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.992610] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "286f24dd-2239-434e-8ae7-15164a0ea8b0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.992884] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "286f24dd-2239-434e-8ae7-15164a0ea8b0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.992983] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "286f24dd-2239-434e-8ae7-15164a0ea8b0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.999031] env[61987]: INFO nova.compute.manager [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Terminating instance [ 952.005552] env[61987]: DEBUG nova.compute.provider_tree [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.042590] env[61987]: DEBUG nova.network.neutron [-] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.071372] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5d2b608d-89cd-4af6-a1f2-7eff47165ca4 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "865222ae-12da-4e5a-84af-041003b50520" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.537s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.323744] env[61987]: DEBUG nova.compute.manager [req-8cec9315-514d-459c-aa40-aea7714ceb59 req-3cb6085a-b8e8-45c2-827d-5bed1f671795 service nova] [instance: 865222ae-12da-4e5a-84af-041003b50520] Received event network-changed-9df062f4-8440-4694-9896-000aa92195e2 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 952.323951] env[61987]: DEBUG nova.compute.manager [req-8cec9315-514d-459c-aa40-aea7714ceb59 req-3cb6085a-b8e8-45c2-827d-5bed1f671795 service nova] [instance: 865222ae-12da-4e5a-84af-041003b50520] Refreshing instance network info cache due to event network-changed-9df062f4-8440-4694-9896-000aa92195e2. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 952.324257] env[61987]: DEBUG oslo_concurrency.lockutils [req-8cec9315-514d-459c-aa40-aea7714ceb59 req-3cb6085a-b8e8-45c2-827d-5bed1f671795 service nova] Acquiring lock "refresh_cache-865222ae-12da-4e5a-84af-041003b50520" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.324432] env[61987]: DEBUG oslo_concurrency.lockutils [req-8cec9315-514d-459c-aa40-aea7714ceb59 req-3cb6085a-b8e8-45c2-827d-5bed1f671795 service nova] Acquired lock "refresh_cache-865222ae-12da-4e5a-84af-041003b50520" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.324602] env[61987]: DEBUG nova.network.neutron [req-8cec9315-514d-459c-aa40-aea7714ceb59 req-3cb6085a-b8e8-45c2-827d-5bed1f671795 service nova] [instance: 865222ae-12da-4e5a-84af-041003b50520] Refreshing network info cache for port 9df062f4-8440-4694-9896-000aa92195e2 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 952.508389] env[61987]: DEBUG nova.scheduler.client.report [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 952.516212] env[61987]: DEBUG nova.compute.manager [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 952.516212] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 952.516918] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f09c1646-b7de-45ac-9460-05dab402f87e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.522439] env[61987]: DEBUG nova.compute.manager [req-44b74010-d5e9-4b6c-a13a-0dae793bf210 req-19e3cd11-38a5-4573-b122-7407a4fd1f7f service nova] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Received event network-vif-deleted-0e240bad-236d-411b-b538-f0025aa3c0ea {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 952.527981] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 952.527981] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3c553986-8447-4d41-83ad-c5dea13c0ecd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.534670] env[61987]: DEBUG oslo_vmware.api [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 952.534670] env[61987]: value = "task-1062032" [ 952.534670] env[61987]: _type = "Task" [ 952.534670] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.545493] env[61987]: INFO nova.compute.manager [-] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Took 2.01 seconds to deallocate network for instance. [ 952.545860] env[61987]: DEBUG oslo_vmware.api [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1062032, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.018794] env[61987]: DEBUG oslo_concurrency.lockutils [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.343s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.019547] env[61987]: DEBUG nova.compute.manager [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 953.027746] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.253s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.028065] env[61987]: DEBUG nova.objects.instance [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lazy-loading 'resources' on Instance uuid a720083b-7b16-468a-a595-bfe8592e8621 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.048732] env[61987]: DEBUG oslo_vmware.api [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1062032, 'name': PowerOffVM_Task, 'duration_secs': 0.268122} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.049595] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 953.049595] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 953.049595] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-64903adc-8cfc-4509-9571-8a9cffebf849 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.053811] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.113691] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 953.113945] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 953.114177] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Deleting the datastore file [datastore2] 286f24dd-2239-434e-8ae7-15164a0ea8b0 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 953.114890] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-458f1d00-713d-4cd6-ae41-905c1041add1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.126542] env[61987]: DEBUG oslo_vmware.api [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 953.126542] env[61987]: value = "task-1062034" [ 953.126542] env[61987]: _type = "Task" [ 953.126542] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.139284] env[61987]: DEBUG oslo_vmware.api [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1062034, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.193075] env[61987]: DEBUG nova.network.neutron [req-8cec9315-514d-459c-aa40-aea7714ceb59 req-3cb6085a-b8e8-45c2-827d-5bed1f671795 service nova] [instance: 865222ae-12da-4e5a-84af-041003b50520] Updated VIF entry in instance network info cache for port 9df062f4-8440-4694-9896-000aa92195e2. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 953.193483] env[61987]: DEBUG nova.network.neutron [req-8cec9315-514d-459c-aa40-aea7714ceb59 req-3cb6085a-b8e8-45c2-827d-5bed1f671795 service nova] [instance: 865222ae-12da-4e5a-84af-041003b50520] Updating instance_info_cache with network_info: [{"id": "9df062f4-8440-4694-9896-000aa92195e2", "address": "fa:16:3e:15:06:44", "network": {"id": "26d453d1-8edf-45e1-9e49-2a3225de6633", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1400564876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b77fbbfc7f6437083f9f30e8fdeeb35", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9df062f4-84", "ovs_interfaceid": "9df062f4-8440-4694-9896-000aa92195e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.531624] env[61987]: DEBUG nova.compute.utils [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 953.535934] env[61987]: DEBUG nova.compute.manager [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 953.536169] env[61987]: DEBUG nova.network.neutron [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 953.605433] env[61987]: DEBUG nova.policy [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '026a788d5fd24553bb62d6ca1c83b36a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fb3548425b6e42ecbcace02a38eeb899', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 953.642206] env[61987]: DEBUG oslo_vmware.api [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1062034, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166071} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.645121] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 953.645335] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 953.645520] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 953.645705] env[61987]: INFO nova.compute.manager [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Took 1.13 seconds to destroy the instance on the hypervisor. [ 953.646016] env[61987]: DEBUG oslo.service.loopingcall [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.646389] env[61987]: DEBUG nova.compute.manager [-] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 953.646482] env[61987]: DEBUG nova.network.neutron [-] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 953.701523] env[61987]: DEBUG oslo_concurrency.lockutils [req-8cec9315-514d-459c-aa40-aea7714ceb59 req-3cb6085a-b8e8-45c2-827d-5bed1f671795 service nova] Releasing lock "refresh_cache-865222ae-12da-4e5a-84af-041003b50520" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.760049] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ef0034-8bbf-4d20-820b-2d6b6e661cde {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.768513] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea20966a-b0bd-420b-a834-1cdd72ab4480 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.804876] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb6f6e0-23e2-494b-aa6f-2ef379c522a9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.813176] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-175cdabd-2cf9-4244-964f-8c63e8e5031b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.829830] env[61987]: DEBUG nova.compute.provider_tree [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 954.037118] env[61987]: DEBUG nova.compute.manager [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 954.044231] env[61987]: DEBUG nova.network.neutron [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Successfully created port: 8ad3014f-98be-4e30-8250-f19043c4b103 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 954.333946] env[61987]: DEBUG nova.scheduler.client.report [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 954.352590] env[61987]: DEBUG nova.compute.manager [req-e4c527e4-6d2e-4420-ae5f-7b089bf54602 req-7e7a0ba8-fab8-474b-b0d8-82f876290d78 service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Received event network-changed-feaf0c72-3782-4054-907b-f98cc72cd02a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 954.352839] env[61987]: DEBUG nova.compute.manager [req-e4c527e4-6d2e-4420-ae5f-7b089bf54602 req-7e7a0ba8-fab8-474b-b0d8-82f876290d78 service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Refreshing instance network info cache due to event network-changed-feaf0c72-3782-4054-907b-f98cc72cd02a. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 954.355412] env[61987]: DEBUG oslo_concurrency.lockutils [req-e4c527e4-6d2e-4420-ae5f-7b089bf54602 req-7e7a0ba8-fab8-474b-b0d8-82f876290d78 service nova] Acquiring lock "refresh_cache-cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.355412] env[61987]: DEBUG oslo_concurrency.lockutils [req-e4c527e4-6d2e-4420-ae5f-7b089bf54602 req-7e7a0ba8-fab8-474b-b0d8-82f876290d78 service nova] Acquired lock "refresh_cache-cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.355412] env[61987]: DEBUG nova.network.neutron [req-e4c527e4-6d2e-4420-ae5f-7b089bf54602 req-7e7a0ba8-fab8-474b-b0d8-82f876290d78 service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Refreshing network info cache for port feaf0c72-3782-4054-907b-f98cc72cd02a {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 954.449826] env[61987]: DEBUG nova.network.neutron [-] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.555890] env[61987]: DEBUG oslo_concurrency.lockutils [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "4d15848d-5265-4897-bfa7-f3ad4c02ed83" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.556232] env[61987]: DEBUG oslo_concurrency.lockutils [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "4d15848d-5265-4897-bfa7-f3ad4c02ed83" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.839302] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.811s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.842229] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.907s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.842725] env[61987]: DEBUG nova.objects.instance [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lazy-loading 'resources' on Instance uuid f9c82f38-7e06-4949-80a5-6eaa3c01c88c {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.858658] env[61987]: INFO nova.scheduler.client.report [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Deleted allocations for instance a720083b-7b16-468a-a595-bfe8592e8621 [ 954.953039] env[61987]: INFO nova.compute.manager [-] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Took 1.31 seconds to deallocate network for instance. [ 955.049776] env[61987]: DEBUG nova.compute.manager [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 955.057889] env[61987]: DEBUG nova.compute.manager [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 955.083239] env[61987]: DEBUG nova.virt.hardware [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 955.083554] env[61987]: DEBUG nova.virt.hardware [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 955.083724] env[61987]: DEBUG nova.virt.hardware [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 955.083919] env[61987]: DEBUG nova.virt.hardware [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 955.084088] env[61987]: DEBUG nova.virt.hardware [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 955.084249] env[61987]: DEBUG nova.virt.hardware [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 955.084464] env[61987]: DEBUG nova.virt.hardware [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 955.084632] env[61987]: DEBUG nova.virt.hardware [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 955.084885] env[61987]: DEBUG nova.virt.hardware [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 955.085074] env[61987]: DEBUG nova.virt.hardware [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 955.085662] env[61987]: DEBUG nova.virt.hardware [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 955.086856] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b720c671-d434-4c4c-86a3-6d0815f85510 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.090328] env[61987]: DEBUG nova.network.neutron [req-e4c527e4-6d2e-4420-ae5f-7b089bf54602 req-7e7a0ba8-fab8-474b-b0d8-82f876290d78 service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Updated VIF entry in instance network info cache for port feaf0c72-3782-4054-907b-f98cc72cd02a. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 955.090675] env[61987]: DEBUG nova.network.neutron [req-e4c527e4-6d2e-4420-ae5f-7b089bf54602 req-7e7a0ba8-fab8-474b-b0d8-82f876290d78 service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Updating instance_info_cache with network_info: [{"id": "feaf0c72-3782-4054-907b-f98cc72cd02a", "address": "fa:16:3e:13:f5:64", "network": {"id": "b45b790f-b724-4a9a-b54b-8f8319b13a88", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1186325730-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7f68eb92e124eab9beb7b60ddcc1df3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfeaf0c72-37", "ovs_interfaceid": "feaf0c72-3782-4054-907b-f98cc72cd02a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.099038] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-245ae940-e004-4c22-9372-c8c272208bd1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.368561] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9fcf58f3-09e6-4181-9e83-1ce6ee7c381c tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "a720083b-7b16-468a-a595-bfe8592e8621" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.580s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.440424] env[61987]: DEBUG oslo_concurrency.lockutils [None req-19c48f3f-dca8-46f9-a6d4-08c45ca516fe tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "interface-efdf62cc-9a3f-4b6b-a119-29a8096848ee-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.440526] env[61987]: DEBUG oslo_concurrency.lockutils [None req-19c48f3f-dca8-46f9-a6d4-08c45ca516fe tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "interface-efdf62cc-9a3f-4b6b-a119-29a8096848ee-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.440871] env[61987]: DEBUG nova.objects.instance [None req-19c48f3f-dca8-46f9-a6d4-08c45ca516fe tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lazy-loading 'flavor' on Instance uuid efdf62cc-9a3f-4b6b-a119-29a8096848ee {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.463018] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.546350] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2565cb31-90cf-4a51-8c01-3cc9f799878f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.554698] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42f2d7e7-640f-4f7e-971f-51414c0be9f9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.588148] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d98186-f9fe-45b5-8616-3369e4696b4f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.593579] env[61987]: DEBUG oslo_concurrency.lockutils [req-e4c527e4-6d2e-4420-ae5f-7b089bf54602 req-7e7a0ba8-fab8-474b-b0d8-82f876290d78 service nova] Releasing lock "refresh_cache-cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.594183] env[61987]: DEBUG nova.compute.manager [req-e4c527e4-6d2e-4420-ae5f-7b089bf54602 req-7e7a0ba8-fab8-474b-b0d8-82f876290d78 service nova] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Received event network-vif-deleted-dfa5785c-1dca-49b8-88ed-a067c0dfb815 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 955.594455] env[61987]: INFO nova.compute.manager [req-e4c527e4-6d2e-4420-ae5f-7b089bf54602 req-7e7a0ba8-fab8-474b-b0d8-82f876290d78 service nova] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Neutron deleted interface dfa5785c-1dca-49b8-88ed-a067c0dfb815; detaching it from the instance and deleting it from the info cache [ 955.594754] env[61987]: DEBUG nova.network.neutron [req-e4c527e4-6d2e-4420-ae5f-7b089bf54602 req-7e7a0ba8-fab8-474b-b0d8-82f876290d78 service nova] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.599427] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ced4daa-e5dc-4bca-a851-b4a964bdc52a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.618118] env[61987]: DEBUG nova.compute.provider_tree [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 955.620701] env[61987]: DEBUG oslo_concurrency.lockutils [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.703334] env[61987]: DEBUG nova.network.neutron [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Successfully updated port: 8ad3014f-98be-4e30-8250-f19043c4b103 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 955.879229] env[61987]: DEBUG nova.compute.manager [req-a9bf7d85-441a-4bd8-848f-2aa8a1ace059 req-f18c772f-a2bb-45c4-8505-51b75e449ba6 service nova] [instance: 70702eab-b510-4946-a426-92902b591481] Received event network-vif-plugged-8ad3014f-98be-4e30-8250-f19043c4b103 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 955.879229] env[61987]: DEBUG oslo_concurrency.lockutils [req-a9bf7d85-441a-4bd8-848f-2aa8a1ace059 req-f18c772f-a2bb-45c4-8505-51b75e449ba6 service nova] Acquiring lock "70702eab-b510-4946-a426-92902b591481-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.879428] env[61987]: DEBUG oslo_concurrency.lockutils [req-a9bf7d85-441a-4bd8-848f-2aa8a1ace059 req-f18c772f-a2bb-45c4-8505-51b75e449ba6 service nova] Lock "70702eab-b510-4946-a426-92902b591481-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.879584] env[61987]: DEBUG oslo_concurrency.lockutils [req-a9bf7d85-441a-4bd8-848f-2aa8a1ace059 req-f18c772f-a2bb-45c4-8505-51b75e449ba6 service nova] Lock "70702eab-b510-4946-a426-92902b591481-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.879810] env[61987]: DEBUG nova.compute.manager [req-a9bf7d85-441a-4bd8-848f-2aa8a1ace059 req-f18c772f-a2bb-45c4-8505-51b75e449ba6 service nova] [instance: 70702eab-b510-4946-a426-92902b591481] No waiting events found dispatching network-vif-plugged-8ad3014f-98be-4e30-8250-f19043c4b103 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 955.880017] env[61987]: WARNING nova.compute.manager [req-a9bf7d85-441a-4bd8-848f-2aa8a1ace059 req-f18c772f-a2bb-45c4-8505-51b75e449ba6 service nova] [instance: 70702eab-b510-4946-a426-92902b591481] Received unexpected event network-vif-plugged-8ad3014f-98be-4e30-8250-f19043c4b103 for instance with vm_state building and task_state spawning. [ 955.948403] env[61987]: DEBUG nova.objects.instance [None req-19c48f3f-dca8-46f9-a6d4-08c45ca516fe tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lazy-loading 'pci_requests' on Instance uuid efdf62cc-9a3f-4b6b-a119-29a8096848ee {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 956.107168] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f4fd33fa-3ddb-493c-bdb3-a9ae85285ee7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.116415] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d3aa6e6-f855-496f-83f5-a8a8a26aa038 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.128241] env[61987]: DEBUG nova.scheduler.client.report [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 956.150265] env[61987]: DEBUG nova.compute.manager [req-e4c527e4-6d2e-4420-ae5f-7b089bf54602 req-7e7a0ba8-fab8-474b-b0d8-82f876290d78 service nova] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Detach interface failed, port_id=dfa5785c-1dca-49b8-88ed-a067c0dfb815, reason: Instance 286f24dd-2239-434e-8ae7-15164a0ea8b0 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 956.207412] env[61987]: DEBUG oslo_concurrency.lockutils [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Acquiring lock "refresh_cache-70702eab-b510-4946-a426-92902b591481" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.207567] env[61987]: DEBUG oslo_concurrency.lockutils [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Acquired lock "refresh_cache-70702eab-b510-4946-a426-92902b591481" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.207718] env[61987]: DEBUG nova.network.neutron [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 956.448901] env[61987]: DEBUG oslo_concurrency.lockutils [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "4dc624ca-c23c-4c4f-8481-45f954e1c068" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.449154] env[61987]: DEBUG oslo_concurrency.lockutils [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "4dc624ca-c23c-4c4f-8481-45f954e1c068" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.453194] env[61987]: DEBUG nova.objects.base [None req-19c48f3f-dca8-46f9-a6d4-08c45ca516fe tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61987) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 956.453392] env[61987]: DEBUG nova.network.neutron [None req-19c48f3f-dca8-46f9-a6d4-08c45ca516fe tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 956.580669] env[61987]: DEBUG oslo_concurrency.lockutils [None req-19c48f3f-dca8-46f9-a6d4-08c45ca516fe tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "interface-efdf62cc-9a3f-4b6b-a119-29a8096848ee-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.140s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.632358] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.791s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.634710] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.581s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.634966] env[61987]: DEBUG nova.objects.instance [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Lazy-loading 'resources' on Instance uuid d11104e7-4a5c-44c3-bd92-d36c587da794 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 956.648953] env[61987]: INFO nova.scheduler.client.report [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Deleted allocations for instance f9c82f38-7e06-4949-80a5-6eaa3c01c88c [ 956.752045] env[61987]: DEBUG nova.network.neutron [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 956.902823] env[61987]: DEBUG nova.network.neutron [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Updating instance_info_cache with network_info: [{"id": "8ad3014f-98be-4e30-8250-f19043c4b103", "address": "fa:16:3e:43:67:7d", "network": {"id": "6248b7c0-2154-4f6b-9f25-3cec510f1a3e", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-2097108713-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb3548425b6e42ecbcace02a38eeb899", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ad3014f-98", "ovs_interfaceid": "8ad3014f-98be-4e30-8250-f19043c4b103", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.951423] env[61987]: DEBUG nova.compute.manager [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 957.158345] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c6e3e15e-81cd-4b95-ad73-6c28ff802dcb tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "f9c82f38-7e06-4949-80a5-6eaa3c01c88c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.670s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.291236] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30b81954-4599-4bc6-abb8-ec190cb26229 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.298804] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e3b28e-dd0b-4d2d-9665-6ed777f66df4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.327938] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e73deb-964f-433e-9997-dc1053011c18 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.335316] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67a8b71-c2c4-4c8e-9fd8-7cb81fef57aa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.348364] env[61987]: DEBUG nova.compute.provider_tree [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.407060] env[61987]: DEBUG oslo_concurrency.lockutils [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Releasing lock "refresh_cache-70702eab-b510-4946-a426-92902b591481" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.407436] env[61987]: DEBUG nova.compute.manager [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Instance network_info: |[{"id": "8ad3014f-98be-4e30-8250-f19043c4b103", "address": "fa:16:3e:43:67:7d", "network": {"id": "6248b7c0-2154-4f6b-9f25-3cec510f1a3e", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-2097108713-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb3548425b6e42ecbcace02a38eeb899", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ad3014f-98", "ovs_interfaceid": "8ad3014f-98be-4e30-8250-f19043c4b103", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 957.407876] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:43:67:7d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd7d0d95-6848-4e69-ac21-75f8db82a3b5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8ad3014f-98be-4e30-8250-f19043c4b103', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 957.415613] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Creating folder: Project (fb3548425b6e42ecbcace02a38eeb899). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 957.415938] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4d0ee6d4-a5a8-45de-afa3-0b99f63eda03 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.427267] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Created folder: Project (fb3548425b6e42ecbcace02a38eeb899) in parent group-v234219. [ 957.427527] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Creating folder: Instances. Parent ref: group-v234395. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 957.427778] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ce432dbd-7748-45cf-96c6-52541092e66f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.436696] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Created folder: Instances in parent group-v234395. [ 957.437596] env[61987]: DEBUG oslo.service.loopingcall [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.437596] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 70702eab-b510-4946-a426-92902b591481] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 957.437596] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9dee0365-f93a-4bbb-8b10-b0349bf371eb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.464199] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 957.464199] env[61987]: value = "task-1062037" [ 957.464199] env[61987]: _type = "Task" [ 957.464199] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.469826] env[61987]: DEBUG oslo_concurrency.lockutils [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.472806] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062037, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.529098] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "1b08af4a-8e3e-405e-8526-1e0a37b1f0cd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.529379] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "1b08af4a-8e3e-405e-8526-1e0a37b1f0cd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.529600] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "1b08af4a-8e3e-405e-8526-1e0a37b1f0cd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.529799] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "1b08af4a-8e3e-405e-8526-1e0a37b1f0cd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.530027] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "1b08af4a-8e3e-405e-8526-1e0a37b1f0cd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.532329] env[61987]: INFO nova.compute.manager [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Terminating instance [ 957.851823] env[61987]: DEBUG nova.scheduler.client.report [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 957.903339] env[61987]: DEBUG nova.compute.manager [req-ed70b04e-f8f6-46f4-8c47-eecbd448bf67 req-acc2cb94-1b1d-4b8d-9df2-17196bdeb552 service nova] [instance: 70702eab-b510-4946-a426-92902b591481] Received event network-changed-8ad3014f-98be-4e30-8250-f19043c4b103 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 957.903552] env[61987]: DEBUG nova.compute.manager [req-ed70b04e-f8f6-46f4-8c47-eecbd448bf67 req-acc2cb94-1b1d-4b8d-9df2-17196bdeb552 service nova] [instance: 70702eab-b510-4946-a426-92902b591481] Refreshing instance network info cache due to event network-changed-8ad3014f-98be-4e30-8250-f19043c4b103. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 957.903780] env[61987]: DEBUG oslo_concurrency.lockutils [req-ed70b04e-f8f6-46f4-8c47-eecbd448bf67 req-acc2cb94-1b1d-4b8d-9df2-17196bdeb552 service nova] Acquiring lock "refresh_cache-70702eab-b510-4946-a426-92902b591481" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.903933] env[61987]: DEBUG oslo_concurrency.lockutils [req-ed70b04e-f8f6-46f4-8c47-eecbd448bf67 req-acc2cb94-1b1d-4b8d-9df2-17196bdeb552 service nova] Acquired lock "refresh_cache-70702eab-b510-4946-a426-92902b591481" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.904116] env[61987]: DEBUG nova.network.neutron [req-ed70b04e-f8f6-46f4-8c47-eecbd448bf67 req-acc2cb94-1b1d-4b8d-9df2-17196bdeb552 service nova] [instance: 70702eab-b510-4946-a426-92902b591481] Refreshing network info cache for port 8ad3014f-98be-4e30-8250-f19043c4b103 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 957.975728] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062037, 'name': CreateVM_Task, 'duration_secs': 0.37404} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.975958] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 70702eab-b510-4946-a426-92902b591481] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 957.976708] env[61987]: DEBUG oslo_concurrency.lockutils [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.976878] env[61987]: DEBUG oslo_concurrency.lockutils [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.977252] env[61987]: DEBUG oslo_concurrency.lockutils [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 957.977639] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-556775b3-2a11-459c-ba5b-485ff4b44230 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.987192] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Waiting for the task: (returnval){ [ 957.987192] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]523a6497-98e5-c689-599f-accfd2df061f" [ 957.987192] env[61987]: _type = "Task" [ 957.987192] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.996584] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523a6497-98e5-c689-599f-accfd2df061f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.036263] env[61987]: DEBUG nova.compute.manager [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 958.036494] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 958.037366] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-281f5ab1-6828-497b-8f97-6a66ad1eb353 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.044759] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 958.045018] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d2e87f36-7e30-4658-a3f4-549a246cb365 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.050452] env[61987]: DEBUG oslo_vmware.api [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 958.050452] env[61987]: value = "task-1062038" [ 958.050452] env[61987]: _type = "Task" [ 958.050452] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.059127] env[61987]: DEBUG oslo_vmware.api [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062038, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.357495] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.723s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.359912] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.900s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.360181] env[61987]: DEBUG nova.objects.instance [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lazy-loading 'resources' on Instance uuid 286f24dd-2239-434e-8ae7-15164a0ea8b0 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.384541] env[61987]: INFO nova.scheduler.client.report [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Deleted allocations for instance d11104e7-4a5c-44c3-bd92-d36c587da794 [ 958.485240] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "interface-efdf62cc-9a3f-4b6b-a119-29a8096848ee-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.485493] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "interface-efdf62cc-9a3f-4b6b-a119-29a8096848ee-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.485964] env[61987]: DEBUG nova.objects.instance [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lazy-loading 'flavor' on Instance uuid efdf62cc-9a3f-4b6b-a119-29a8096848ee {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.500020] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523a6497-98e5-c689-599f-accfd2df061f, 'name': SearchDatastore_Task, 'duration_secs': 0.014978} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.500020] env[61987]: DEBUG oslo_concurrency.lockutils [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.500020] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 958.500020] env[61987]: DEBUG oslo_concurrency.lockutils [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.500020] env[61987]: DEBUG oslo_concurrency.lockutils [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.500020] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 958.500020] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dc13480f-5a1f-4b84-b365-115e68bf8dbf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.507733] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 958.508248] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 958.508703] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05c89a52-a85b-41dd-bd03-0554b4b5c9a8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.513666] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Waiting for the task: (returnval){ [ 958.513666] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52176fb8-98e7-fcc8-7e37-4b78fae4dd7a" [ 958.513666] env[61987]: _type = "Task" [ 958.513666] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.521887] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52176fb8-98e7-fcc8-7e37-4b78fae4dd7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.561975] env[61987]: DEBUG oslo_vmware.api [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062038, 'name': PowerOffVM_Task, 'duration_secs': 0.288608} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.562278] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 958.562455] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 958.562715] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0a6980d4-608f-4255-90fe-c37788bd3369 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.680730] env[61987]: DEBUG nova.network.neutron [req-ed70b04e-f8f6-46f4-8c47-eecbd448bf67 req-acc2cb94-1b1d-4b8d-9df2-17196bdeb552 service nova] [instance: 70702eab-b510-4946-a426-92902b591481] Updated VIF entry in instance network info cache for port 8ad3014f-98be-4e30-8250-f19043c4b103. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 958.681146] env[61987]: DEBUG nova.network.neutron [req-ed70b04e-f8f6-46f4-8c47-eecbd448bf67 req-acc2cb94-1b1d-4b8d-9df2-17196bdeb552 service nova] [instance: 70702eab-b510-4946-a426-92902b591481] Updating instance_info_cache with network_info: [{"id": "8ad3014f-98be-4e30-8250-f19043c4b103", "address": "fa:16:3e:43:67:7d", "network": {"id": "6248b7c0-2154-4f6b-9f25-3cec510f1a3e", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-2097108713-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fb3548425b6e42ecbcace02a38eeb899", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ad3014f-98", "ovs_interfaceid": "8ad3014f-98be-4e30-8250-f19043c4b103", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.893606] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b02e1a98-73b2-4240-897a-b054c35f2e0f tempest-ServersTestJSON-1099866935 tempest-ServersTestJSON-1099866935-project-member] Lock "d11104e7-4a5c-44c3-bd92-d36c587da794" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.014s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.026206] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52176fb8-98e7-fcc8-7e37-4b78fae4dd7a, 'name': SearchDatastore_Task, 'duration_secs': 0.009513} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.029986] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1f06ba1-2010-4e10-b483-cbcd569ae6f9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.032580] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73c23060-987a-4aea-9538-158a58af82b4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.038341] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Waiting for the task: (returnval){ [ 959.038341] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5209a612-5bae-8be5-f81b-a397291f0195" [ 959.038341] env[61987]: _type = "Task" [ 959.038341] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.043710] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d39535f3-ffa1-4400-954a-a5486467355a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.052981] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5209a612-5bae-8be5-f81b-a397291f0195, 'name': SearchDatastore_Task, 'duration_secs': 0.009097} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.077282] env[61987]: DEBUG oslo_concurrency.lockutils [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.077609] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 70702eab-b510-4946-a426-92902b591481/70702eab-b510-4946-a426-92902b591481.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 959.078883] env[61987]: DEBUG nova.objects.instance [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lazy-loading 'pci_requests' on Instance uuid efdf62cc-9a3f-4b6b-a119-29a8096848ee {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.080019] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f01a61f0-b39c-4636-ba6f-f1b70b29369a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.082685] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7bf23b0-2d1c-4f00-a619-c52290e1dcca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.095021] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53f93d59-687f-4f8a-bd98-432427daccaa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.096654] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Waiting for the task: (returnval){ [ 959.096654] env[61987]: value = "task-1062040" [ 959.096654] env[61987]: _type = "Task" [ 959.096654] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.110232] env[61987]: DEBUG nova.compute.provider_tree [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.116323] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Task: {'id': task-1062040, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.184587] env[61987]: DEBUG oslo_concurrency.lockutils [req-ed70b04e-f8f6-46f4-8c47-eecbd448bf67 req-acc2cb94-1b1d-4b8d-9df2-17196bdeb552 service nova] Releasing lock "refresh_cache-70702eab-b510-4946-a426-92902b591481" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.587186] env[61987]: DEBUG nova.objects.base [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61987) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 959.589650] env[61987]: DEBUG nova.network.neutron [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 959.608201] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Task: {'id': task-1062040, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.613653] env[61987]: DEBUG nova.scheduler.client.report [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 959.637812] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 959.638054] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 959.638273] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Deleting the datastore file [datastore1] 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 959.638786] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-804df394-865e-4685-b632-3a28f9eb7594 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.646081] env[61987]: DEBUG oslo_vmware.api [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 959.646081] env[61987]: value = "task-1062041" [ 959.646081] env[61987]: _type = "Task" [ 959.646081] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.654138] env[61987]: DEBUG oslo_vmware.api [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062041, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.655694] env[61987]: DEBUG nova.policy [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '697f9b1e3b3840369549a4f1ae3173ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd2afdc9bddba4ccab2b8fdb1f31da2b6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 959.939255] env[61987]: DEBUG nova.network.neutron [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Successfully created port: 697d9636-5a3c-4b4e-8a86-3ea919b04757 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 960.109643] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Task: {'id': task-1062040, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.960966} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.109939] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 70702eab-b510-4946-a426-92902b591481/70702eab-b510-4946-a426-92902b591481.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 960.110188] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 960.110460] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c5a44dcd-894f-4bcd-9e12-97d0df0102c8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.117010] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Waiting for the task: (returnval){ [ 960.117010] env[61987]: value = "task-1062042" [ 960.117010] env[61987]: _type = "Task" [ 960.117010] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.121311] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.761s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.123533] env[61987]: DEBUG oslo_concurrency.lockutils [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.503s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.128160] env[61987]: INFO nova.compute.claims [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 960.133351] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Task: {'id': task-1062042, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.162777] env[61987]: DEBUG oslo_vmware.api [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062041, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.33999} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.164086] env[61987]: INFO nova.scheduler.client.report [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Deleted allocations for instance 286f24dd-2239-434e-8ae7-15164a0ea8b0 [ 960.166915] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 960.166915] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 960.169273] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 960.169273] env[61987]: INFO nova.compute.manager [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Took 2.13 seconds to destroy the instance on the hypervisor. [ 960.169273] env[61987]: DEBUG oslo.service.loopingcall [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 960.172758] env[61987]: DEBUG nova.compute.manager [-] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 960.172758] env[61987]: DEBUG nova.network.neutron [-] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 960.427131] env[61987]: DEBUG nova.compute.manager [req-afbc0673-358e-4afb-9f56-a255943de4f2 req-f81eb65b-56ba-4ff5-826f-0cf0a719ad24 service nova] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Received event network-vif-deleted-09512bd5-f20a-49d6-aaf6-d4fa46cb0889 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 960.427263] env[61987]: INFO nova.compute.manager [req-afbc0673-358e-4afb-9f56-a255943de4f2 req-f81eb65b-56ba-4ff5-826f-0cf0a719ad24 service nova] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Neutron deleted interface 09512bd5-f20a-49d6-aaf6-d4fa46cb0889; detaching it from the instance and deleting it from the info cache [ 960.427460] env[61987]: DEBUG nova.network.neutron [req-afbc0673-358e-4afb-9f56-a255943de4f2 req-f81eb65b-56ba-4ff5-826f-0cf0a719ad24 service nova] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.627897] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Task: {'id': task-1062042, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081671} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.628223] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 960.629037] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d727c37-f7a4-47fe-b6e5-d43e8286a442 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.658686] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 70702eab-b510-4946-a426-92902b591481/70702eab-b510-4946-a426-92902b591481.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 960.659730] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a3c25b0-cf00-4356-b8ed-45a54b09c99d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.678144] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0883fca6-727c-4307-87b0-1d48c53c9459 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "286f24dd-2239-434e-8ae7-15164a0ea8b0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.685s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.682747] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Waiting for the task: (returnval){ [ 960.682747] env[61987]: value = "task-1062043" [ 960.682747] env[61987]: _type = "Task" [ 960.682747] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.692930] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Task: {'id': task-1062043, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.906203] env[61987]: DEBUG nova.network.neutron [-] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.930938] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3f282731-8a13-46bd-ab82-1656b91f1f18 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.941764] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46e9fdb6-f944-4ac1-8f11-ec7c1960754f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.976404] env[61987]: DEBUG nova.compute.manager [req-afbc0673-358e-4afb-9f56-a255943de4f2 req-f81eb65b-56ba-4ff5-826f-0cf0a719ad24 service nova] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Detach interface failed, port_id=09512bd5-f20a-49d6-aaf6-d4fa46cb0889, reason: Instance 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 961.191697] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Task: {'id': task-1062043, 'name': ReconfigVM_Task, 'duration_secs': 0.305566} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.194075] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 70702eab-b510-4946-a426-92902b591481/70702eab-b510-4946-a426-92902b591481.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 961.194854] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dbca6151-8f25-4c94-9e93-dfa2c1722bb5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.200993] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Waiting for the task: (returnval){ [ 961.200993] env[61987]: value = "task-1062044" [ 961.200993] env[61987]: _type = "Task" [ 961.200993] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.209705] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Task: {'id': task-1062044, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.284894] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b96e99db-a190-4858-95d5-bc4ea28e724c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.292306] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e671ab-5885-4475-912d-1a0587bb13f9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.323930] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2632c627-23bd-42c4-a222-fbdffa54327e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.331127] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f20e5e70-16bc-4a60-acce-3f9e5d7ff848 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.344251] env[61987]: DEBUG nova.compute.provider_tree [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.408858] env[61987]: DEBUG oslo_concurrency.lockutils [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.409163] env[61987]: DEBUG oslo_concurrency.lockutils [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.409378] env[61987]: DEBUG oslo_concurrency.lockutils [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.409567] env[61987]: DEBUG oslo_concurrency.lockutils [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.409746] env[61987]: DEBUG oslo_concurrency.lockutils [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.411771] env[61987]: INFO nova.compute.manager [-] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Took 1.24 seconds to deallocate network for instance. [ 961.416803] env[61987]: INFO nova.compute.manager [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Terminating instance [ 961.503271] env[61987]: DEBUG nova.compute.manager [req-06f7869c-3eff-4f9e-951f-31d06a551ab3 req-baab22d3-d6a2-46b5-a5ba-7b13a791dad1 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Received event network-vif-plugged-697d9636-5a3c-4b4e-8a86-3ea919b04757 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 961.503796] env[61987]: DEBUG oslo_concurrency.lockutils [req-06f7869c-3eff-4f9e-951f-31d06a551ab3 req-baab22d3-d6a2-46b5-a5ba-7b13a791dad1 service nova] Acquiring lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.504045] env[61987]: DEBUG oslo_concurrency.lockutils [req-06f7869c-3eff-4f9e-951f-31d06a551ab3 req-baab22d3-d6a2-46b5-a5ba-7b13a791dad1 service nova] Lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.504236] env[61987]: DEBUG oslo_concurrency.lockutils [req-06f7869c-3eff-4f9e-951f-31d06a551ab3 req-baab22d3-d6a2-46b5-a5ba-7b13a791dad1 service nova] Lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.504447] env[61987]: DEBUG nova.compute.manager [req-06f7869c-3eff-4f9e-951f-31d06a551ab3 req-baab22d3-d6a2-46b5-a5ba-7b13a791dad1 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] No waiting events found dispatching network-vif-plugged-697d9636-5a3c-4b4e-8a86-3ea919b04757 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 961.504632] env[61987]: WARNING nova.compute.manager [req-06f7869c-3eff-4f9e-951f-31d06a551ab3 req-baab22d3-d6a2-46b5-a5ba-7b13a791dad1 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Received unexpected event network-vif-plugged-697d9636-5a3c-4b4e-8a86-3ea919b04757 for instance with vm_state active and task_state None. [ 961.590871] env[61987]: DEBUG nova.network.neutron [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Successfully updated port: 697d9636-5a3c-4b4e-8a86-3ea919b04757 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 961.712080] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Task: {'id': task-1062044, 'name': Rename_Task, 'duration_secs': 0.149764} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.712080] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 961.712080] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4805d3d4-a4fb-4989-979c-33cdba7371f6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.717904] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Waiting for the task: (returnval){ [ 961.717904] env[61987]: value = "task-1062045" [ 961.717904] env[61987]: _type = "Task" [ 961.717904] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.724988] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Task: {'id': task-1062045, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.847436] env[61987]: DEBUG nova.scheduler.client.report [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 961.918551] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.920457] env[61987]: DEBUG nova.compute.manager [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 961.920626] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 961.921639] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79320551-c478-4bdb-9b6a-7c01b6da86ef {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.930280] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 961.930570] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ce7f9c36-4c8a-4846-b7b8-b4524abe1579 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.936977] env[61987]: DEBUG oslo_vmware.api [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 961.936977] env[61987]: value = "task-1062046" [ 961.936977] env[61987]: _type = "Task" [ 961.936977] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.945709] env[61987]: DEBUG oslo_vmware.api [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1062046, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.095984] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.095984] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.095984] env[61987]: DEBUG nova.network.neutron [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 962.228527] env[61987]: DEBUG oslo_vmware.api [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Task: {'id': task-1062045, 'name': PowerOnVM_Task, 'duration_secs': 0.464701} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.228827] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 962.229084] env[61987]: INFO nova.compute.manager [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Took 7.18 seconds to spawn the instance on the hypervisor. [ 962.229295] env[61987]: DEBUG nova.compute.manager [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 962.230081] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e5a4d4c-2f72-43a8-8e17-4e628ae54075 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.353178] env[61987]: DEBUG oslo_concurrency.lockutils [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.229s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.353684] env[61987]: DEBUG nova.compute.manager [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 962.356943] env[61987]: DEBUG oslo_concurrency.lockutils [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.887s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.358378] env[61987]: INFO nova.compute.claims [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 962.446515] env[61987]: DEBUG oslo_vmware.api [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1062046, 'name': PowerOffVM_Task, 'duration_secs': 0.208127} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.446809] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 962.446973] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 962.447238] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f9775b7b-3138-4e2c-9f61-eb87b247138c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.510052] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 962.510319] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 962.510508] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Deleting the datastore file [datastore1] 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 962.510786] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f92e40be-a4d4-4d12-880f-cda1b3e16ca3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.517136] env[61987]: DEBUG oslo_vmware.api [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for the task: (returnval){ [ 962.517136] env[61987]: value = "task-1062048" [ 962.517136] env[61987]: _type = "Task" [ 962.517136] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.524484] env[61987]: DEBUG oslo_vmware.api [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1062048, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.632929] env[61987]: WARNING nova.network.neutron [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] 257456d2-c9db-4303-bdfd-035bbb123901 already exists in list: networks containing: ['257456d2-c9db-4303-bdfd-035bbb123901']. ignoring it [ 962.750056] env[61987]: INFO nova.compute.manager [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Took 12.40 seconds to build instance. [ 962.860650] env[61987]: DEBUG nova.compute.utils [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 962.862415] env[61987]: DEBUG nova.compute.manager [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 962.862415] env[61987]: DEBUG nova.network.neutron [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 962.902406] env[61987]: DEBUG nova.policy [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c5cf7a7673cc4f5aadca7960c355c394', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ea5c65ac3b143178b13731663dbd8ee', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 962.930893] env[61987]: DEBUG nova.network.neutron [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Updating instance_info_cache with network_info: [{"id": "c4833bbc-333e-46c2-a140-5e8d493d3973", "address": "fa:16:3e:11:e5:35", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4833bbc-33", "ovs_interfaceid": "c4833bbc-333e-46c2-a140-5e8d493d3973", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "697d9636-5a3c-4b4e-8a86-3ea919b04757", "address": "fa:16:3e:73:76:c3", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap697d9636-5a", "ovs_interfaceid": "697d9636-5a3c-4b4e-8a86-3ea919b04757", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.034414] env[61987]: DEBUG oslo_vmware.api [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Task: {'id': task-1062048, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175256} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.034769] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 963.034769] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 963.034859] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 963.035030] env[61987]: INFO nova.compute.manager [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Took 1.11 seconds to destroy the instance on the hypervisor. [ 963.035286] env[61987]: DEBUG oslo.service.loopingcall [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 963.035491] env[61987]: DEBUG nova.compute.manager [-] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 963.035587] env[61987]: DEBUG nova.network.neutron [-] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 963.185211] env[61987]: DEBUG nova.network.neutron [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Successfully created port: 203d81e4-dbb8-4685-a463-9de48b835ff7 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 963.252574] env[61987]: DEBUG oslo_concurrency.lockutils [None req-601e68a2-95bf-485f-b4f9-8e4dfa24f8c0 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Lock "70702eab-b510-4946-a426-92902b591481" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.909s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.365878] env[61987]: DEBUG nova.compute.manager [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 963.434065] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.434797] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.434994] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.436065] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15632fb6-be0e-4834-8cbf-89b21b988239 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.459742] env[61987]: DEBUG nova.virt.hardware [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 963.460017] env[61987]: DEBUG nova.virt.hardware [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 963.460207] env[61987]: DEBUG nova.virt.hardware [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 963.460407] env[61987]: DEBUG nova.virt.hardware [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 963.460567] env[61987]: DEBUG nova.virt.hardware [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 963.460725] env[61987]: DEBUG nova.virt.hardware [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 963.461257] env[61987]: DEBUG nova.virt.hardware [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 963.461257] env[61987]: DEBUG nova.virt.hardware [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 963.461390] env[61987]: DEBUG nova.virt.hardware [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 963.461546] env[61987]: DEBUG nova.virt.hardware [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 963.461728] env[61987]: DEBUG nova.virt.hardware [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 963.468831] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Reconfiguring VM to attach interface {{(pid=61987) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 963.472213] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4fe006f1-8570-4c83-a0a2-4e9c8b047955 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.493783] env[61987]: DEBUG oslo_vmware.api [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 963.493783] env[61987]: value = "task-1062049" [ 963.493783] env[61987]: _type = "Task" [ 963.493783] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.501763] env[61987]: DEBUG oslo_vmware.api [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062049, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.590857] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d9aaf91-18a8-4946-aa3b-713673faf4e0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.598958] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2698b501-69cd-4dfc-b264-813df277cead {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.629646] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-788a207d-ed2d-4252-a4ce-120106ec17b0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.637062] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-979dea73-d094-4675-b122-725113bd64a1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.650373] env[61987]: DEBUG nova.compute.provider_tree [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.771470] env[61987]: DEBUG nova.network.neutron [-] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.003687] env[61987]: DEBUG oslo_vmware.api [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062049, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.079702] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae8b034c-4aa5-49fa-b1a1-fce706354020 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Acquiring lock "interface-70702eab-b510-4946-a426-92902b591481-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.079702] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae8b034c-4aa5-49fa-b1a1-fce706354020 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Lock "interface-70702eab-b510-4946-a426-92902b591481-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.080253] env[61987]: DEBUG nova.objects.instance [None req-ae8b034c-4aa5-49fa-b1a1-fce706354020 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Lazy-loading 'flavor' on Instance uuid 70702eab-b510-4946-a426-92902b591481 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.153416] env[61987]: DEBUG nova.scheduler.client.report [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 964.273978] env[61987]: INFO nova.compute.manager [-] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Took 1.24 seconds to deallocate network for instance. [ 964.378436] env[61987]: DEBUG nova.compute.manager [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 964.455780] env[61987]: DEBUG nova.virt.hardware [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 964.456084] env[61987]: DEBUG nova.virt.hardware [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 964.456275] env[61987]: DEBUG nova.virt.hardware [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 964.456478] env[61987]: DEBUG nova.virt.hardware [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 964.456636] env[61987]: DEBUG nova.virt.hardware [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 964.456793] env[61987]: DEBUG nova.virt.hardware [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 964.457028] env[61987]: DEBUG nova.virt.hardware [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 964.457233] env[61987]: DEBUG nova.virt.hardware [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 964.457417] env[61987]: DEBUG nova.virt.hardware [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 964.457589] env[61987]: DEBUG nova.virt.hardware [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 964.457773] env[61987]: DEBUG nova.virt.hardware [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 964.458640] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f6649df-e334-4f50-96f8-2bcfb8933e41 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.466509] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9418df6f-eb9c-4871-8ea6-17a8950f6a9e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.500549] env[61987]: DEBUG oslo_vmware.api [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062049, 'name': ReconfigVM_Task, 'duration_secs': 0.697087} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.501032] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.501265] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Reconfigured VM to attach interface {{(pid=61987) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 964.584058] env[61987]: DEBUG nova.objects.instance [None req-ae8b034c-4aa5-49fa-b1a1-fce706354020 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Lazy-loading 'pci_requests' on Instance uuid 70702eab-b510-4946-a426-92902b591481 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.660393] env[61987]: DEBUG oslo_concurrency.lockutils [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.302s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.660393] env[61987]: DEBUG nova.compute.manager [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 964.663370] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.745s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.663526] env[61987]: DEBUG nova.objects.instance [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lazy-loading 'resources' on Instance uuid 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.781528] env[61987]: DEBUG oslo_concurrency.lockutils [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.839389] env[61987]: DEBUG nova.network.neutron [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Successfully updated port: 203d81e4-dbb8-4685-a463-9de48b835ff7 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 965.006047] env[61987]: DEBUG oslo_concurrency.lockutils [None req-88b09e8c-e643-42fd-9416-e8f561815a94 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "interface-efdf62cc-9a3f-4b6b-a119-29a8096848ee-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.520s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.086807] env[61987]: DEBUG nova.objects.base [None req-ae8b034c-4aa5-49fa-b1a1-fce706354020 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Object Instance<70702eab-b510-4946-a426-92902b591481> lazy-loaded attributes: flavor,pci_requests {{(pid=61987) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 965.087216] env[61987]: DEBUG nova.network.neutron [None req-ae8b034c-4aa5-49fa-b1a1-fce706354020 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 965.161475] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae8b034c-4aa5-49fa-b1a1-fce706354020 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Lock "interface-70702eab-b510-4946-a426-92902b591481-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.082s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.167040] env[61987]: DEBUG nova.compute.utils [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 965.171246] env[61987]: DEBUG nova.compute.manager [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 965.171350] env[61987]: DEBUG nova.network.neutron [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 965.234501] env[61987]: DEBUG nova.policy [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c124c6b758543a68e141796b585e41f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bd47eb44bc334bf3ae5813905903ecbc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 965.322257] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15800eb9-b218-4cc0-a5b7-beda72e1e55c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.329444] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5859a1a-d740-4fb3-862f-dd29e56baa53 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.360561] env[61987]: DEBUG oslo_concurrency.lockutils [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "refresh_cache-4d15848d-5265-4897-bfa7-f3ad4c02ed83" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.360717] env[61987]: DEBUG oslo_concurrency.lockutils [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired lock "refresh_cache-4d15848d-5265-4897-bfa7-f3ad4c02ed83" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.360874] env[61987]: DEBUG nova.network.neutron [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 965.362930] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20719986-886a-408d-a2ba-24aed7d462df {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.370529] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9723e862-9cad-485f-9213-c024568a2e8e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.384882] env[61987]: DEBUG nova.compute.provider_tree [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.514925] env[61987]: DEBUG nova.network.neutron [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Successfully created port: 1735c272-060e-4a5d-836a-af4e2eed480a {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 965.671925] env[61987]: DEBUG nova.compute.manager [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 965.888080] env[61987]: DEBUG nova.scheduler.client.report [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 965.892020] env[61987]: DEBUG nova.network.neutron [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 966.018997] env[61987]: DEBUG nova.network.neutron [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Updating instance_info_cache with network_info: [{"id": "203d81e4-dbb8-4685-a463-9de48b835ff7", "address": "fa:16:3e:5c:c5:2c", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap203d81e4-db", "ovs_interfaceid": "203d81e4-dbb8-4685-a463-9de48b835ff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.089162] env[61987]: DEBUG nova.compute.manager [req-db3e7923-ec88-4466-8efe-403b1ab0fbdf req-4e083b39-8455-4338-b359-c8f6759bf9a7 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Received event network-changed-697d9636-5a3c-4b4e-8a86-3ea919b04757 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 966.089418] env[61987]: DEBUG nova.compute.manager [req-db3e7923-ec88-4466-8efe-403b1ab0fbdf req-4e083b39-8455-4338-b359-c8f6759bf9a7 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Refreshing instance network info cache due to event network-changed-697d9636-5a3c-4b4e-8a86-3ea919b04757. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 966.089560] env[61987]: DEBUG oslo_concurrency.lockutils [req-db3e7923-ec88-4466-8efe-403b1ab0fbdf req-4e083b39-8455-4338-b359-c8f6759bf9a7 service nova] Acquiring lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.089720] env[61987]: DEBUG oslo_concurrency.lockutils [req-db3e7923-ec88-4466-8efe-403b1ab0fbdf req-4e083b39-8455-4338-b359-c8f6759bf9a7 service nova] Acquired lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.089889] env[61987]: DEBUG nova.network.neutron [req-db3e7923-ec88-4466-8efe-403b1ab0fbdf req-4e083b39-8455-4338-b359-c8f6759bf9a7 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Refreshing network info cache for port 697d9636-5a3c-4b4e-8a86-3ea919b04757 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 966.396039] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.733s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.398446] env[61987]: DEBUG oslo_concurrency.lockutils [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.617s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.398736] env[61987]: DEBUG nova.objects.instance [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lazy-loading 'resources' on Instance uuid 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.419417] env[61987]: INFO nova.scheduler.client.report [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Deleted allocations for instance 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd [ 966.450782] env[61987]: DEBUG nova.compute.manager [req-7673e3b0-a357-484c-a62f-51404b5690cd req-5f340463-29e9-425a-85a2-67d76f96120b service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Received event network-vif-plugged-203d81e4-dbb8-4685-a463-9de48b835ff7 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 966.451064] env[61987]: DEBUG oslo_concurrency.lockutils [req-7673e3b0-a357-484c-a62f-51404b5690cd req-5f340463-29e9-425a-85a2-67d76f96120b service nova] Acquiring lock "4d15848d-5265-4897-bfa7-f3ad4c02ed83-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.451324] env[61987]: DEBUG oslo_concurrency.lockutils [req-7673e3b0-a357-484c-a62f-51404b5690cd req-5f340463-29e9-425a-85a2-67d76f96120b service nova] Lock "4d15848d-5265-4897-bfa7-f3ad4c02ed83-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.451583] env[61987]: DEBUG oslo_concurrency.lockutils [req-7673e3b0-a357-484c-a62f-51404b5690cd req-5f340463-29e9-425a-85a2-67d76f96120b service nova] Lock "4d15848d-5265-4897-bfa7-f3ad4c02ed83-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.451822] env[61987]: DEBUG nova.compute.manager [req-7673e3b0-a357-484c-a62f-51404b5690cd req-5f340463-29e9-425a-85a2-67d76f96120b service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] No waiting events found dispatching network-vif-plugged-203d81e4-dbb8-4685-a463-9de48b835ff7 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 966.451954] env[61987]: WARNING nova.compute.manager [req-7673e3b0-a357-484c-a62f-51404b5690cd req-5f340463-29e9-425a-85a2-67d76f96120b service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Received unexpected event network-vif-plugged-203d81e4-dbb8-4685-a463-9de48b835ff7 for instance with vm_state building and task_state spawning. [ 966.521783] env[61987]: DEBUG oslo_concurrency.lockutils [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Releasing lock "refresh_cache-4d15848d-5265-4897-bfa7-f3ad4c02ed83" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.522132] env[61987]: DEBUG nova.compute.manager [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Instance network_info: |[{"id": "203d81e4-dbb8-4685-a463-9de48b835ff7", "address": "fa:16:3e:5c:c5:2c", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap203d81e4-db", "ovs_interfaceid": "203d81e4-dbb8-4685-a463-9de48b835ff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 966.522741] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:c5:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'be8bd197-4b2b-46e7-88ea-2554b0438584', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '203d81e4-dbb8-4685-a463-9de48b835ff7', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 966.530375] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Creating folder: Project (8ea5c65ac3b143178b13731663dbd8ee). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 966.530889] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ff8a2edf-a7b1-4800-9147-5016838dc3fc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.540878] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Created folder: Project (8ea5c65ac3b143178b13731663dbd8ee) in parent group-v234219. [ 966.541160] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Creating folder: Instances. Parent ref: group-v234398. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 966.541365] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aea77067-2718-44da-9b8c-f1519cc4d152 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.549192] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Created folder: Instances in parent group-v234398. [ 966.549427] env[61987]: DEBUG oslo.service.loopingcall [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 966.549620] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 966.549817] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-56cf7c05-4fd7-417b-becb-231ae5d7f7ff {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.569591] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 966.569591] env[61987]: value = "task-1062052" [ 966.569591] env[61987]: _type = "Task" [ 966.569591] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.577962] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062052, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.681181] env[61987]: DEBUG nova.compute.manager [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 966.713104] env[61987]: DEBUG nova.virt.hardware [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 966.713377] env[61987]: DEBUG nova.virt.hardware [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 966.713555] env[61987]: DEBUG nova.virt.hardware [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 966.713750] env[61987]: DEBUG nova.virt.hardware [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 966.713904] env[61987]: DEBUG nova.virt.hardware [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 966.714455] env[61987]: DEBUG nova.virt.hardware [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 966.714796] env[61987]: DEBUG nova.virt.hardware [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 966.715010] env[61987]: DEBUG nova.virt.hardware [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 966.715743] env[61987]: DEBUG nova.virt.hardware [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 966.715743] env[61987]: DEBUG nova.virt.hardware [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 966.715743] env[61987]: DEBUG nova.virt.hardware [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 966.716849] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b8e2a80-8964-4acb-83c2-2aacc968c177 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.727112] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-427998cb-dbea-4407-9950-3ee7940bd881 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.929092] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7a82579c-3062-416f-8786-8bb4ba4cee28 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "1b08af4a-8e3e-405e-8526-1e0a37b1f0cd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.400s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.000691] env[61987]: DEBUG nova.network.neutron [req-db3e7923-ec88-4466-8efe-403b1ab0fbdf req-4e083b39-8455-4338-b359-c8f6759bf9a7 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Updated VIF entry in instance network info cache for port 697d9636-5a3c-4b4e-8a86-3ea919b04757. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 967.001153] env[61987]: DEBUG nova.network.neutron [req-db3e7923-ec88-4466-8efe-403b1ab0fbdf req-4e083b39-8455-4338-b359-c8f6759bf9a7 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Updating instance_info_cache with network_info: [{"id": "c4833bbc-333e-46c2-a140-5e8d493d3973", "address": "fa:16:3e:11:e5:35", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4833bbc-33", "ovs_interfaceid": "c4833bbc-333e-46c2-a140-5e8d493d3973", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "697d9636-5a3c-4b4e-8a86-3ea919b04757", "address": "fa:16:3e:73:76:c3", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap697d9636-5a", "ovs_interfaceid": "697d9636-5a3c-4b4e-8a86-3ea919b04757", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.059129] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9af322-e0b2-44c7-80e9-a29fd79759db {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.066565] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3786887-3f82-4e65-90b0-b2243ce41577 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.104655] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f13400f1-3506-45bb-b563-56497f1c87f2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.109824] env[61987]: DEBUG nova.network.neutron [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Successfully updated port: 1735c272-060e-4a5d-836a-af4e2eed480a {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 967.111231] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062052, 'name': CreateVM_Task, 'duration_secs': 0.364876} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.112633] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 967.117733] env[61987]: DEBUG oslo_concurrency.lockutils [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.117733] env[61987]: DEBUG oslo_concurrency.lockutils [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.117733] env[61987]: DEBUG oslo_concurrency.lockutils [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 967.118610] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12af5642-a0f7-4294-8f89-6c7c85d923c5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.121134] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3952e5b6-50b2-45e1-a030-7b939ccc582c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.128683] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 967.128683] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52ad023b-52da-4ded-8a6f-f510b14d409c" [ 967.128683] env[61987]: _type = "Task" [ 967.128683] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.136743] env[61987]: DEBUG nova.compute.provider_tree [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.145726] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52ad023b-52da-4ded-8a6f-f510b14d409c, 'name': SearchDatastore_Task, 'duration_secs': 0.009085} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.146411] env[61987]: DEBUG oslo_concurrency.lockutils [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.146641] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 967.146880] env[61987]: DEBUG oslo_concurrency.lockutils [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.147048] env[61987]: DEBUG oslo_concurrency.lockutils [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.147233] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 967.147671] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-49fbed5f-3cc4-4846-b6ff-1c0dea15d8d1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.154480] env[61987]: DEBUG oslo_concurrency.lockutils [None req-613db91b-8097-4dad-9845-0dbe9b8e6464 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "69a22e40-d469-4500-926e-0a12a233f252" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.154665] env[61987]: DEBUG oslo_concurrency.lockutils [None req-613db91b-8097-4dad-9845-0dbe9b8e6464 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "69a22e40-d469-4500-926e-0a12a233f252" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.154891] env[61987]: INFO nova.compute.manager [None req-613db91b-8097-4dad-9845-0dbe9b8e6464 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Rebooting instance [ 967.156112] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 967.156295] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 967.157322] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a0ad47e-a392-4ace-9aa2-5152039b337d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.162384] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 967.162384] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52351647-175c-c294-b40f-55831d3a5d08" [ 967.162384] env[61987]: _type = "Task" [ 967.162384] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.163191] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Acquiring lock "70702eab-b510-4946-a426-92902b591481" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.163404] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Lock "70702eab-b510-4946-a426-92902b591481" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.163621] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Acquiring lock "70702eab-b510-4946-a426-92902b591481-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.163810] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Lock "70702eab-b510-4946-a426-92902b591481-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.163988] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Lock "70702eab-b510-4946-a426-92902b591481-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.168464] env[61987]: INFO nova.compute.manager [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Terminating instance [ 967.174827] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52351647-175c-c294-b40f-55831d3a5d08, 'name': SearchDatastore_Task, 'duration_secs': 0.008281} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.175761] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9ff1a0a-309c-4fbe-ad18-4cbc03a8b1df {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.180864] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 967.180864] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]522f7b51-84fc-738c-1eb2-7a172d8bb996" [ 967.180864] env[61987]: _type = "Task" [ 967.180864] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.189680] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]522f7b51-84fc-738c-1eb2-7a172d8bb996, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.503805] env[61987]: DEBUG oslo_concurrency.lockutils [req-db3e7923-ec88-4466-8efe-403b1ab0fbdf req-4e083b39-8455-4338-b359-c8f6759bf9a7 service nova] Releasing lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.504124] env[61987]: DEBUG nova.compute.manager [req-db3e7923-ec88-4466-8efe-403b1ab0fbdf req-4e083b39-8455-4338-b359-c8f6759bf9a7 service nova] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Received event network-vif-deleted-a60eaa2f-cdd3-4b07-9a40-d48a4dd95d17 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 967.612469] env[61987]: DEBUG oslo_concurrency.lockutils [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "refresh_cache-4dc624ca-c23c-4c4f-8481-45f954e1c068" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.612721] env[61987]: DEBUG oslo_concurrency.lockutils [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired lock "refresh_cache-4dc624ca-c23c-4c4f-8481-45f954e1c068" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.612853] env[61987]: DEBUG nova.network.neutron [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 967.619093] env[61987]: DEBUG oslo_concurrency.lockutils [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "interface-efdf62cc-9a3f-4b6b-a119-29a8096848ee-b42049a4-5eda-4603-950e-36103e3b012a" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.619372] env[61987]: DEBUG oslo_concurrency.lockutils [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "interface-efdf62cc-9a3f-4b6b-a119-29a8096848ee-b42049a4-5eda-4603-950e-36103e3b012a" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.619716] env[61987]: DEBUG nova.objects.instance [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lazy-loading 'flavor' on Instance uuid efdf62cc-9a3f-4b6b-a119-29a8096848ee {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.642675] env[61987]: DEBUG nova.scheduler.client.report [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 967.672121] env[61987]: DEBUG nova.compute.manager [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 967.672306] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 967.673738] env[61987]: DEBUG oslo_concurrency.lockutils [None req-613db91b-8097-4dad-9845-0dbe9b8e6464 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.673738] env[61987]: DEBUG oslo_concurrency.lockutils [None req-613db91b-8097-4dad-9845-0dbe9b8e6464 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.673888] env[61987]: DEBUG nova.network.neutron [None req-613db91b-8097-4dad-9845-0dbe9b8e6464 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 967.675638] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630da780-a4b7-41ff-a9e9-7730309ce893 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.686707] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 967.687763] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fa583489-23f2-46e2-9dd1-b250730055d0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.693401] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]522f7b51-84fc-738c-1eb2-7a172d8bb996, 'name': SearchDatastore_Task, 'duration_secs': 0.009007} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.693650] env[61987]: DEBUG oslo_concurrency.lockutils [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.693914] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 4d15848d-5265-4897-bfa7-f3ad4c02ed83/4d15848d-5265-4897-bfa7-f3ad4c02ed83.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 967.694173] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5879c186-af33-4da5-8fbe-87933002f0aa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.699222] env[61987]: DEBUG oslo_vmware.api [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Waiting for the task: (returnval){ [ 967.699222] env[61987]: value = "task-1062053" [ 967.699222] env[61987]: _type = "Task" [ 967.699222] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.702801] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 967.702801] env[61987]: value = "task-1062054" [ 967.702801] env[61987]: _type = "Task" [ 967.702801] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.708897] env[61987]: DEBUG oslo_vmware.api [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Task: {'id': task-1062053, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.713493] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062054, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.132110] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "6ec74ee9-a73b-4eac-97f2-abac3b1e3191" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.132110] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "6ec74ee9-a73b-4eac-97f2-abac3b1e3191" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.147099] env[61987]: DEBUG oslo_concurrency.lockutils [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.749s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.150113] env[61987]: DEBUG nova.network.neutron [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 968.210306] env[61987]: DEBUG oslo_vmware.api [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Task: {'id': task-1062053, 'name': PowerOffVM_Task, 'duration_secs': 0.190839} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.210991] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 968.211204] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 968.211470] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7c1027a4-558c-4e63-8f8c-938709242982 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.216086] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062054, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487535} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.218841] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 4d15848d-5265-4897-bfa7-f3ad4c02ed83/4d15848d-5265-4897-bfa7-f3ad4c02ed83.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 968.219087] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 968.219348] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-785c1756-8869-4c42-8705-21794f8636f0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.226066] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 968.226066] env[61987]: value = "task-1062056" [ 968.226066] env[61987]: _type = "Task" [ 968.226066] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.233909] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062056, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.253870] env[61987]: INFO nova.scheduler.client.report [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Deleted allocations for instance 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32 [ 968.289149] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 968.289409] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 968.289598] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Deleting the datastore file [datastore2] 70702eab-b510-4946-a426-92902b591481 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.289876] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bbebf2a1-65d6-44c4-86b8-0d8b63cab013 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.296372] env[61987]: DEBUG oslo_vmware.api [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Waiting for the task: (returnval){ [ 968.296372] env[61987]: value = "task-1062057" [ 968.296372] env[61987]: _type = "Task" [ 968.296372] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.305844] env[61987]: DEBUG oslo_vmware.api [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Task: {'id': task-1062057, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.306740] env[61987]: DEBUG nova.network.neutron [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Updating instance_info_cache with network_info: [{"id": "1735c272-060e-4a5d-836a-af4e2eed480a", "address": "fa:16:3e:69:d8:56", "network": {"id": "ca68bf96-3f18-45ee-80fb-64a1260ed996", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-733595377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd47eb44bc334bf3ae5813905903ecbc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a27fd90b-16a5-43af-bede-ae36762ece00", "external-id": "nsx-vlan-transportzone-197", "segmentation_id": 197, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1735c272-06", "ovs_interfaceid": "1735c272-060e-4a5d-836a-af4e2eed480a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.478677] env[61987]: DEBUG nova.compute.manager [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Received event network-changed-203d81e4-dbb8-4685-a463-9de48b835ff7 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 968.478883] env[61987]: DEBUG nova.compute.manager [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Refreshing instance network info cache due to event network-changed-203d81e4-dbb8-4685-a463-9de48b835ff7. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 968.479134] env[61987]: DEBUG oslo_concurrency.lockutils [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] Acquiring lock "refresh_cache-4d15848d-5265-4897-bfa7-f3ad4c02ed83" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.479289] env[61987]: DEBUG oslo_concurrency.lockutils [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] Acquired lock "refresh_cache-4d15848d-5265-4897-bfa7-f3ad4c02ed83" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.479455] env[61987]: DEBUG nova.network.neutron [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Refreshing network info cache for port 203d81e4-dbb8-4685-a463-9de48b835ff7 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 968.483649] env[61987]: DEBUG nova.objects.instance [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lazy-loading 'pci_requests' on Instance uuid efdf62cc-9a3f-4b6b-a119-29a8096848ee {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 968.625317] env[61987]: DEBUG nova.network.neutron [None req-613db91b-8097-4dad-9845-0dbe9b8e6464 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updating instance_info_cache with network_info: [{"id": "785edefc-5e84-4a34-97e1-b7c1a08f218e", "address": "fa:16:3e:bf:8b:d7", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap785edefc-5e", "ovs_interfaceid": "785edefc-5e84-4a34-97e1-b7c1a08f218e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.634563] env[61987]: DEBUG nova.compute.manager [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 968.736315] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062056, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068762} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.736520] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 968.737365] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1072ad31-7f35-4e4e-952f-f1ff0e38a36c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.761765] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] 4d15848d-5265-4897-bfa7-f3ad4c02ed83/4d15848d-5265-4897-bfa7-f3ad4c02ed83.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 968.764337] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-936c5e22-4c7b-4692-94c9-42ee32189f21 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.779172] env[61987]: DEBUG oslo_concurrency.lockutils [None req-31716f36-c06d-4cb4-9bd6-e248c53d7821 tempest-ListServerFiltersTestJSON-357176646 tempest-ListServerFiltersTestJSON-357176646-project-member] Lock "5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.370s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.786432] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 968.786432] env[61987]: value = "task-1062058" [ 968.786432] env[61987]: _type = "Task" [ 968.786432] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.794606] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062058, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.805101] env[61987]: DEBUG oslo_vmware.api [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Task: {'id': task-1062057, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166747} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.805365] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 968.806059] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 968.806059] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 968.806059] env[61987]: INFO nova.compute.manager [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] [instance: 70702eab-b510-4946-a426-92902b591481] Took 1.13 seconds to destroy the instance on the hypervisor. [ 968.806339] env[61987]: DEBUG oslo.service.loopingcall [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 968.806339] env[61987]: DEBUG nova.compute.manager [-] [instance: 70702eab-b510-4946-a426-92902b591481] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 968.806447] env[61987]: DEBUG nova.network.neutron [-] [instance: 70702eab-b510-4946-a426-92902b591481] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 968.808565] env[61987]: DEBUG oslo_concurrency.lockutils [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Releasing lock "refresh_cache-4dc624ca-c23c-4c4f-8481-45f954e1c068" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.808854] env[61987]: DEBUG nova.compute.manager [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Instance network_info: |[{"id": "1735c272-060e-4a5d-836a-af4e2eed480a", "address": "fa:16:3e:69:d8:56", "network": {"id": "ca68bf96-3f18-45ee-80fb-64a1260ed996", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-733595377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd47eb44bc334bf3ae5813905903ecbc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a27fd90b-16a5-43af-bede-ae36762ece00", "external-id": "nsx-vlan-transportzone-197", "segmentation_id": 197, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1735c272-06", "ovs_interfaceid": "1735c272-060e-4a5d-836a-af4e2eed480a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 968.809237] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:69:d8:56', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a27fd90b-16a5-43af-bede-ae36762ece00', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1735c272-060e-4a5d-836a-af4e2eed480a', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 968.816833] env[61987]: DEBUG oslo.service.loopingcall [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 968.817329] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 968.817548] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ffe6f41b-e598-47ed-89db-b4068267ad57 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.836762] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 968.836762] env[61987]: value = "task-1062059" [ 968.836762] env[61987]: _type = "Task" [ 968.836762] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.844436] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062059, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.985600] env[61987]: DEBUG nova.objects.base [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61987) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 968.986112] env[61987]: DEBUG nova.network.neutron [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 969.102087] env[61987]: DEBUG nova.policy [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '697f9b1e3b3840369549a4f1ae3173ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd2afdc9bddba4ccab2b8fdb1f31da2b6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 969.127931] env[61987]: DEBUG oslo_concurrency.lockutils [None req-613db91b-8097-4dad-9845-0dbe9b8e6464 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.159484] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.159799] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.161169] env[61987]: INFO nova.compute.claims [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 969.296533] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062058, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.350676] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062059, 'name': CreateVM_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.404557] env[61987]: DEBUG nova.network.neutron [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Updated VIF entry in instance network info cache for port 203d81e4-dbb8-4685-a463-9de48b835ff7. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 969.405011] env[61987]: DEBUG nova.network.neutron [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Updating instance_info_cache with network_info: [{"id": "203d81e4-dbb8-4685-a463-9de48b835ff7", "address": "fa:16:3e:5c:c5:2c", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap203d81e4-db", "ovs_interfaceid": "203d81e4-dbb8-4685-a463-9de48b835ff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.430992] env[61987]: DEBUG nova.compute.manager [req-6d852d88-e072-43e1-a096-2701edb0e542 req-ac178f88-71e4-41d9-bf7d-b9d0d7f94db3 service nova] [instance: 70702eab-b510-4946-a426-92902b591481] Received event network-vif-deleted-8ad3014f-98be-4e30-8250-f19043c4b103 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 969.430992] env[61987]: INFO nova.compute.manager [req-6d852d88-e072-43e1-a096-2701edb0e542 req-ac178f88-71e4-41d9-bf7d-b9d0d7f94db3 service nova] [instance: 70702eab-b510-4946-a426-92902b591481] Neutron deleted interface 8ad3014f-98be-4e30-8250-f19043c4b103; detaching it from the instance and deleting it from the info cache [ 969.430992] env[61987]: DEBUG nova.network.neutron [req-6d852d88-e072-43e1-a096-2701edb0e542 req-ac178f88-71e4-41d9-bf7d-b9d0d7f94db3 service nova] [instance: 70702eab-b510-4946-a426-92902b591481] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.632731] env[61987]: DEBUG nova.compute.manager [None req-613db91b-8097-4dad-9845-0dbe9b8e6464 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 969.633692] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f62fd6-b8bd-4fc5-ad39-cfe6235f59e1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.796371] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062058, 'name': ReconfigVM_Task, 'duration_secs': 0.61812} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.796671] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Reconfigured VM instance instance-00000060 to attach disk [datastore2] 4d15848d-5265-4897-bfa7-f3ad4c02ed83/4d15848d-5265-4897-bfa7-f3ad4c02ed83.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 969.797380] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-58a6ea29-a136-4326-b3ef-d41385ab77eb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.798963] env[61987]: DEBUG nova.network.neutron [-] [instance: 70702eab-b510-4946-a426-92902b591481] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.805377] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 969.805377] env[61987]: value = "task-1062060" [ 969.805377] env[61987]: _type = "Task" [ 969.805377] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.814439] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062060, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.847417] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062059, 'name': CreateVM_Task, 'duration_secs': 0.764243} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.847600] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 969.848306] env[61987]: DEBUG oslo_concurrency.lockutils [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.848487] env[61987]: DEBUG oslo_concurrency.lockutils [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.848826] env[61987]: DEBUG oslo_concurrency.lockutils [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 969.849106] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0cd98d72-cc1c-47b2-8955-8a771144773d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.853856] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 969.853856] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52a0fdad-3940-d225-84fd-2a3d570b6123" [ 969.853856] env[61987]: _type = "Task" [ 969.853856] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.861508] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52a0fdad-3940-d225-84fd-2a3d570b6123, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.908197] env[61987]: DEBUG oslo_concurrency.lockutils [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] Releasing lock "refresh_cache-4d15848d-5265-4897-bfa7-f3ad4c02ed83" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.908485] env[61987]: DEBUG nova.compute.manager [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Received event network-vif-plugged-1735c272-060e-4a5d-836a-af4e2eed480a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 969.908718] env[61987]: DEBUG oslo_concurrency.lockutils [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] Acquiring lock "4dc624ca-c23c-4c4f-8481-45f954e1c068-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.908990] env[61987]: DEBUG oslo_concurrency.lockutils [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] Lock "4dc624ca-c23c-4c4f-8481-45f954e1c068-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.909277] env[61987]: DEBUG oslo_concurrency.lockutils [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] Lock "4dc624ca-c23c-4c4f-8481-45f954e1c068-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.909534] env[61987]: DEBUG nova.compute.manager [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] No waiting events found dispatching network-vif-plugged-1735c272-060e-4a5d-836a-af4e2eed480a {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 969.909733] env[61987]: WARNING nova.compute.manager [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Received unexpected event network-vif-plugged-1735c272-060e-4a5d-836a-af4e2eed480a for instance with vm_state building and task_state spawning. [ 969.909938] env[61987]: DEBUG nova.compute.manager [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Received event network-changed-1735c272-060e-4a5d-836a-af4e2eed480a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 969.910146] env[61987]: DEBUG nova.compute.manager [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Refreshing instance network info cache due to event network-changed-1735c272-060e-4a5d-836a-af4e2eed480a. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 969.910360] env[61987]: DEBUG oslo_concurrency.lockutils [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] Acquiring lock "refresh_cache-4dc624ca-c23c-4c4f-8481-45f954e1c068" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.910508] env[61987]: DEBUG oslo_concurrency.lockutils [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] Acquired lock "refresh_cache-4dc624ca-c23c-4c4f-8481-45f954e1c068" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.910697] env[61987]: DEBUG nova.network.neutron [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Refreshing network info cache for port 1735c272-060e-4a5d-836a-af4e2eed480a {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 969.932728] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-28ff3dc8-912c-41c7-82c4-5e64cb7d81c3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.942281] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf767870-545d-4513-8516-e0a0c530b711 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.970451] env[61987]: DEBUG nova.compute.manager [req-6d852d88-e072-43e1-a096-2701edb0e542 req-ac178f88-71e4-41d9-bf7d-b9d0d7f94db3 service nova] [instance: 70702eab-b510-4946-a426-92902b591481] Detach interface failed, port_id=8ad3014f-98be-4e30-8250-f19043c4b103, reason: Instance 70702eab-b510-4946-a426-92902b591481 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 970.301034] env[61987]: INFO nova.compute.manager [-] [instance: 70702eab-b510-4946-a426-92902b591481] Took 1.49 seconds to deallocate network for instance. [ 970.317333] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062060, 'name': Rename_Task, 'duration_secs': 0.173999} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.318493] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 970.319240] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1706181b-fe55-4589-9886-564d5ab45019 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.321727] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-937ef1a6-758d-4dcf-b0fa-f6404d27650e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.328707] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c6c235-8614-4e00-aa58-677bff406a9f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.331888] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 970.331888] env[61987]: value = "task-1062061" [ 970.331888] env[61987]: _type = "Task" [ 970.331888] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.364264] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c32888-2d38-4473-acfe-c0c9c88b98d3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.371174] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062061, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.376627] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52a0fdad-3940-d225-84fd-2a3d570b6123, 'name': SearchDatastore_Task, 'duration_secs': 0.008975} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.378749] env[61987]: DEBUG oslo_concurrency.lockutils [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.378998] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 970.379265] env[61987]: DEBUG oslo_concurrency.lockutils [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.379566] env[61987]: DEBUG oslo_concurrency.lockutils [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.379651] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 970.379890] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-322bffa0-464f-4854-a01f-6b5d4e9d8291 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.382524] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02e74c62-df68-42c9-8543-137fd5848301 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.398491] env[61987]: DEBUG nova.compute.provider_tree [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.401588] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 970.401588] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 970.404022] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20ec73ea-2fc4-4b83-8588-4e3bbd48e15c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.407195] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 970.407195] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52db1340-934c-2fab-9768-075bb4c9b4ef" [ 970.407195] env[61987]: _type = "Task" [ 970.407195] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.417958] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52db1340-934c-2fab-9768-075bb4c9b4ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.653280] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db5cd6fc-5e3a-47ae-aaed-02e8b28d2de5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.662765] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-613db91b-8097-4dad-9845-0dbe9b8e6464 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Doing hard reboot of VM {{(pid=61987) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 970.664034] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-04cb54a8-3a89-4fdf-9555-88a07d2e3d7d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.670129] env[61987]: DEBUG oslo_vmware.api [None req-613db91b-8097-4dad-9845-0dbe9b8e6464 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 970.670129] env[61987]: value = "task-1062062" [ 970.670129] env[61987]: _type = "Task" [ 970.670129] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.670996] env[61987]: DEBUG nova.network.neutron [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Updated VIF entry in instance network info cache for port 1735c272-060e-4a5d-836a-af4e2eed480a. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 970.671364] env[61987]: DEBUG nova.network.neutron [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Updating instance_info_cache with network_info: [{"id": "1735c272-060e-4a5d-836a-af4e2eed480a", "address": "fa:16:3e:69:d8:56", "network": {"id": "ca68bf96-3f18-45ee-80fb-64a1260ed996", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-733595377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd47eb44bc334bf3ae5813905903ecbc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a27fd90b-16a5-43af-bede-ae36762ece00", "external-id": "nsx-vlan-transportzone-197", "segmentation_id": 197, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1735c272-06", "ovs_interfaceid": "1735c272-060e-4a5d-836a-af4e2eed480a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.681984] env[61987]: DEBUG oslo_vmware.api [None req-613db91b-8097-4dad-9845-0dbe9b8e6464 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062062, 'name': ResetVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.812715] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.842230] env[61987]: DEBUG oslo_vmware.api [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062061, 'name': PowerOnVM_Task, 'duration_secs': 0.450419} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.842485] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 970.842729] env[61987]: INFO nova.compute.manager [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Took 6.46 seconds to spawn the instance on the hypervisor. [ 970.842934] env[61987]: DEBUG nova.compute.manager [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 970.843688] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53f61d0b-cd77-4c65-a071-12563126bd19 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.901777] env[61987]: DEBUG nova.scheduler.client.report [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 970.919795] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52db1340-934c-2fab-9768-075bb4c9b4ef, 'name': SearchDatastore_Task, 'duration_secs': 0.014536} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.921637] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31de63be-0b9a-48b9-a595-9feceae3aabc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.928635] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 970.928635] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52688fe5-edbb-5e2a-43d3-59d8d2aa1587" [ 970.928635] env[61987]: _type = "Task" [ 970.928635] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.937678] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52688fe5-edbb-5e2a-43d3-59d8d2aa1587, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.993586] env[61987]: DEBUG nova.compute.manager [req-1bb01360-f957-4efd-9a4b-2e5bc41554dc req-152c485d-7517-449b-b45a-a5977c49212a service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Received event network-vif-plugged-b42049a4-5eda-4603-950e-36103e3b012a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 970.993814] env[61987]: DEBUG oslo_concurrency.lockutils [req-1bb01360-f957-4efd-9a4b-2e5bc41554dc req-152c485d-7517-449b-b45a-a5977c49212a service nova] Acquiring lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.994459] env[61987]: DEBUG oslo_concurrency.lockutils [req-1bb01360-f957-4efd-9a4b-2e5bc41554dc req-152c485d-7517-449b-b45a-a5977c49212a service nova] Lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.994773] env[61987]: DEBUG oslo_concurrency.lockutils [req-1bb01360-f957-4efd-9a4b-2e5bc41554dc req-152c485d-7517-449b-b45a-a5977c49212a service nova] Lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.994962] env[61987]: DEBUG nova.compute.manager [req-1bb01360-f957-4efd-9a4b-2e5bc41554dc req-152c485d-7517-449b-b45a-a5977c49212a service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] No waiting events found dispatching network-vif-plugged-b42049a4-5eda-4603-950e-36103e3b012a {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 970.995196] env[61987]: WARNING nova.compute.manager [req-1bb01360-f957-4efd-9a4b-2e5bc41554dc req-152c485d-7517-449b-b45a-a5977c49212a service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Received unexpected event network-vif-plugged-b42049a4-5eda-4603-950e-36103e3b012a for instance with vm_state active and task_state None. [ 971.173730] env[61987]: DEBUG nova.network.neutron [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Successfully updated port: b42049a4-5eda-4603-950e-36103e3b012a {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 971.182934] env[61987]: DEBUG oslo_concurrency.lockutils [req-702d77de-81b3-4889-a30f-04d29a335a35 req-32e666df-6522-4bf8-8ca4-af270a7b80ed service nova] Releasing lock "refresh_cache-4dc624ca-c23c-4c4f-8481-45f954e1c068" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.190272] env[61987]: DEBUG oslo_vmware.api [None req-613db91b-8097-4dad-9845-0dbe9b8e6464 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062062, 'name': ResetVM_Task, 'duration_secs': 0.098602} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.190543] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-613db91b-8097-4dad-9845-0dbe9b8e6464 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Did hard reboot of VM {{(pid=61987) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 971.190752] env[61987]: DEBUG nova.compute.manager [None req-613db91b-8097-4dad-9845-0dbe9b8e6464 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 971.191530] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8bed4e2-19d1-4a17-aa73-8dbebca1cf2b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.363874] env[61987]: INFO nova.compute.manager [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Took 15.78 seconds to build instance. [ 971.407554] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.248s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.408085] env[61987]: DEBUG nova.compute.manager [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 971.412105] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.599s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.412105] env[61987]: DEBUG nova.objects.instance [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Lazy-loading 'resources' on Instance uuid 70702eab-b510-4946-a426-92902b591481 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 971.439854] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52688fe5-edbb-5e2a-43d3-59d8d2aa1587, 'name': SearchDatastore_Task, 'duration_secs': 0.011392} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.440147] env[61987]: DEBUG oslo_concurrency.lockutils [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.440410] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 4dc624ca-c23c-4c4f-8481-45f954e1c068/4dc624ca-c23c-4c4f-8481-45f954e1c068.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 971.441477] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-186ec871-1b07-43ba-9369-91fc0da81ce4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.447886] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 971.447886] env[61987]: value = "task-1062063" [ 971.447886] env[61987]: _type = "Task" [ 971.447886] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.456655] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062063, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.683874] env[61987]: DEBUG oslo_concurrency.lockutils [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.683963] env[61987]: DEBUG oslo_concurrency.lockutils [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.684108] env[61987]: DEBUG nova.network.neutron [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 971.706640] env[61987]: DEBUG oslo_concurrency.lockutils [None req-613db91b-8097-4dad-9845-0dbe9b8e6464 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "69a22e40-d469-4500-926e-0a12a233f252" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.552s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.865958] env[61987]: DEBUG oslo_concurrency.lockutils [None req-68bc84bc-6bc9-489d-a791-b8863963bcc2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "4d15848d-5265-4897-bfa7-f3ad4c02ed83" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.310s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.914897] env[61987]: DEBUG nova.compute.utils [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 971.923628] env[61987]: DEBUG nova.compute.manager [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 971.923628] env[61987]: DEBUG nova.network.neutron [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 971.964392] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062063, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.992389] env[61987]: DEBUG nova.policy [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e16f557dfdb04406b3fc772b891f05a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '22e62741abfb4471bbc7e11ae7d407ca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 972.083982] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-737b10fe-fbb6-485d-85c6-59d192da344e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.092873] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62b57277-22be-4d95-9774-a516139b713a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.129560] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ecf167c-07ff-402a-98f5-315dce4cb6fc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.139515] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cecac73-480d-4f80-b73b-0463d9a7c85f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.155402] env[61987]: DEBUG nova.compute.provider_tree [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.262210] env[61987]: WARNING nova.network.neutron [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] 257456d2-c9db-4303-bdfd-035bbb123901 already exists in list: networks containing: ['257456d2-c9db-4303-bdfd-035bbb123901']. ignoring it [ 972.262670] env[61987]: WARNING nova.network.neutron [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] 257456d2-c9db-4303-bdfd-035bbb123901 already exists in list: networks containing: ['257456d2-c9db-4303-bdfd-035bbb123901']. ignoring it [ 972.420069] env[61987]: DEBUG nova.network.neutron [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Successfully created port: e237b976-f4a7-4dd6-8003-9b80e7f8e5fe {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 972.423039] env[61987]: DEBUG nova.compute.manager [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 972.467041] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062063, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.579596} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.467041] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 4dc624ca-c23c-4c4f-8481-45f954e1c068/4dc624ca-c23c-4c4f-8481-45f954e1c068.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 972.467041] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 972.467802] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bd201f9d-8dfe-42ce-aab7-15dd70d2f007 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.477089] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 972.477089] env[61987]: value = "task-1062064" [ 972.477089] env[61987]: _type = "Task" [ 972.477089] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.488548] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062064, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.661063] env[61987]: DEBUG nova.scheduler.client.report [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 972.987664] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062064, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077782} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.988061] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 972.989202] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70261d7d-ab29-4f9a-90d5-05ad68ca80ea {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.013789] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 4dc624ca-c23c-4c4f-8481-45f954e1c068/4dc624ca-c23c-4c4f-8481-45f954e1c068.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 973.016943] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c31e7495-4eab-410c-810f-385b57509f46 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.040873] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 973.040873] env[61987]: value = "task-1062065" [ 973.040873] env[61987]: _type = "Task" [ 973.040873] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.051155] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062065, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.165376] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.754s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.187264] env[61987]: INFO nova.scheduler.client.report [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Deleted allocations for instance 70702eab-b510-4946-a426-92902b591481 [ 973.320695] env[61987]: DEBUG nova.network.neutron [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Updating instance_info_cache with network_info: [{"id": "c4833bbc-333e-46c2-a140-5e8d493d3973", "address": "fa:16:3e:11:e5:35", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4833bbc-33", "ovs_interfaceid": "c4833bbc-333e-46c2-a140-5e8d493d3973", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "697d9636-5a3c-4b4e-8a86-3ea919b04757", "address": "fa:16:3e:73:76:c3", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap697d9636-5a", "ovs_interfaceid": "697d9636-5a3c-4b4e-8a86-3ea919b04757", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b42049a4-5eda-4603-950e-36103e3b012a", "address": "fa:16:3e:20:21:9e", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb42049a4-5e", "ovs_interfaceid": "b42049a4-5eda-4603-950e-36103e3b012a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.435133] env[61987]: DEBUG nova.compute.manager [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 973.473862] env[61987]: DEBUG nova.virt.hardware [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 973.474250] env[61987]: DEBUG nova.virt.hardware [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 973.474362] env[61987]: DEBUG nova.virt.hardware [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 973.474575] env[61987]: DEBUG nova.virt.hardware [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 973.474914] env[61987]: DEBUG nova.virt.hardware [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 973.474914] env[61987]: DEBUG nova.virt.hardware [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 973.475389] env[61987]: DEBUG nova.virt.hardware [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 973.475606] env[61987]: DEBUG nova.virt.hardware [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 973.475803] env[61987]: DEBUG nova.virt.hardware [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 973.475973] env[61987]: DEBUG nova.virt.hardware [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 973.476274] env[61987]: DEBUG nova.virt.hardware [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 973.477886] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0aa1ebe-b2f5-4b1a-9cd5-e05f49137ebb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.490670] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec891e53-989c-459b-afe8-186e3531e7b5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.508405] env[61987]: DEBUG nova.compute.manager [req-27892a78-01ab-426a-ae7b-3435e38ff1bf req-aaa32319-76b4-4c03-921c-317674f4dde7 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Received event network-changed-b42049a4-5eda-4603-950e-36103e3b012a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 973.508672] env[61987]: DEBUG nova.compute.manager [req-27892a78-01ab-426a-ae7b-3435e38ff1bf req-aaa32319-76b4-4c03-921c-317674f4dde7 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Refreshing instance network info cache due to event network-changed-b42049a4-5eda-4603-950e-36103e3b012a. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 973.508977] env[61987]: DEBUG oslo_concurrency.lockutils [req-27892a78-01ab-426a-ae7b-3435e38ff1bf req-aaa32319-76b4-4c03-921c-317674f4dde7 service nova] Acquiring lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.523166] env[61987]: DEBUG nova.compute.manager [req-5e5d5f53-4d53-4c1f-9964-60c9e2a1686a req-f761e1e3-f7a1-4405-8cd9-f9c04f870f48 service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Received event network-changed-203d81e4-dbb8-4685-a463-9de48b835ff7 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 973.523853] env[61987]: DEBUG nova.compute.manager [req-5e5d5f53-4d53-4c1f-9964-60c9e2a1686a req-f761e1e3-f7a1-4405-8cd9-f9c04f870f48 service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Refreshing instance network info cache due to event network-changed-203d81e4-dbb8-4685-a463-9de48b835ff7. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 973.523853] env[61987]: DEBUG oslo_concurrency.lockutils [req-5e5d5f53-4d53-4c1f-9964-60c9e2a1686a req-f761e1e3-f7a1-4405-8cd9-f9c04f870f48 service nova] Acquiring lock "refresh_cache-4d15848d-5265-4897-bfa7-f3ad4c02ed83" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.524101] env[61987]: DEBUG oslo_concurrency.lockutils [req-5e5d5f53-4d53-4c1f-9964-60c9e2a1686a req-f761e1e3-f7a1-4405-8cd9-f9c04f870f48 service nova] Acquired lock "refresh_cache-4d15848d-5265-4897-bfa7-f3ad4c02ed83" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.524101] env[61987]: DEBUG nova.network.neutron [req-5e5d5f53-4d53-4c1f-9964-60c9e2a1686a req-f761e1e3-f7a1-4405-8cd9-f9c04f870f48 service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Refreshing network info cache for port 203d81e4-dbb8-4685-a463-9de48b835ff7 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 973.554768] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062065, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.699029] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1a7642c5-deb9-4d7d-b5d2-200010c5cd56 tempest-AttachInterfacesV270Test-2133790587 tempest-AttachInterfacesV270Test-2133790587-project-member] Lock "70702eab-b510-4946-a426-92902b591481" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.535s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.825866] env[61987]: DEBUG oslo_concurrency.lockutils [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.826989] env[61987]: DEBUG oslo_concurrency.lockutils [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 973.826989] env[61987]: DEBUG oslo_concurrency.lockutils [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.827154] env[61987]: DEBUG oslo_concurrency.lockutils [req-27892a78-01ab-426a-ae7b-3435e38ff1bf req-aaa32319-76b4-4c03-921c-317674f4dde7 service nova] Acquired lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.827789] env[61987]: DEBUG nova.network.neutron [req-27892a78-01ab-426a-ae7b-3435e38ff1bf req-aaa32319-76b4-4c03-921c-317674f4dde7 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Refreshing network info cache for port b42049a4-5eda-4603-950e-36103e3b012a {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 973.829348] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed40591c-e6e9-458c-bc08-ba65566ca1f9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.852935] env[61987]: DEBUG nova.virt.hardware [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 973.853196] env[61987]: DEBUG nova.virt.hardware [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 973.853388] env[61987]: DEBUG nova.virt.hardware [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 973.853586] env[61987]: DEBUG nova.virt.hardware [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 973.853743] env[61987]: DEBUG nova.virt.hardware [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 973.853900] env[61987]: DEBUG nova.virt.hardware [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 973.854125] env[61987]: DEBUG nova.virt.hardware [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 973.854345] env[61987]: DEBUG nova.virt.hardware [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 973.854530] env[61987]: DEBUG nova.virt.hardware [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 973.854706] env[61987]: DEBUG nova.virt.hardware [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 973.854945] env[61987]: DEBUG nova.virt.hardware [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 973.861226] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Reconfiguring VM to attach interface {{(pid=61987) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 973.861885] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45033785-c801-4fe9-9300-e9d0e30489e5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.884490] env[61987]: DEBUG oslo_vmware.api [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 973.884490] env[61987]: value = "task-1062066" [ 973.884490] env[61987]: _type = "Task" [ 973.884490] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.891185] env[61987]: DEBUG oslo_vmware.api [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062066, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.033169] env[61987]: DEBUG nova.network.neutron [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Successfully updated port: e237b976-f4a7-4dd6-8003-9b80e7f8e5fe {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 974.052826] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062065, 'name': ReconfigVM_Task, 'duration_secs': 0.668329} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.053124] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 4dc624ca-c23c-4c4f-8481-45f954e1c068/4dc624ca-c23c-4c4f-8481-45f954e1c068.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 974.054193] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2dc2f780-5833-41e6-9372-5f03ef120217 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.061032] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 974.061032] env[61987]: value = "task-1062067" [ 974.061032] env[61987]: _type = "Task" [ 974.061032] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.070249] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062067, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.103599] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.103856] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.290539] env[61987]: DEBUG nova.network.neutron [req-5e5d5f53-4d53-4c1f-9964-60c9e2a1686a req-f761e1e3-f7a1-4405-8cd9-f9c04f870f48 service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Updated VIF entry in instance network info cache for port 203d81e4-dbb8-4685-a463-9de48b835ff7. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 974.290962] env[61987]: DEBUG nova.network.neutron [req-5e5d5f53-4d53-4c1f-9964-60c9e2a1686a req-f761e1e3-f7a1-4405-8cd9-f9c04f870f48 service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Updating instance_info_cache with network_info: [{"id": "203d81e4-dbb8-4685-a463-9de48b835ff7", "address": "fa:16:3e:5c:c5:2c", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap203d81e4-db", "ovs_interfaceid": "203d81e4-dbb8-4685-a463-9de48b835ff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.391230] env[61987]: DEBUG oslo_vmware.api [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062066, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.536676] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "refresh_cache-6ec74ee9-a73b-4eac-97f2-abac3b1e3191" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.536676] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired lock "refresh_cache-6ec74ee9-a73b-4eac-97f2-abac3b1e3191" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.537169] env[61987]: DEBUG nova.network.neutron [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 974.545351] env[61987]: DEBUG nova.network.neutron [req-27892a78-01ab-426a-ae7b-3435e38ff1bf req-aaa32319-76b4-4c03-921c-317674f4dde7 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Updated VIF entry in instance network info cache for port b42049a4-5eda-4603-950e-36103e3b012a. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 974.545758] env[61987]: DEBUG nova.network.neutron [req-27892a78-01ab-426a-ae7b-3435e38ff1bf req-aaa32319-76b4-4c03-921c-317674f4dde7 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Updating instance_info_cache with network_info: [{"id": "c4833bbc-333e-46c2-a140-5e8d493d3973", "address": "fa:16:3e:11:e5:35", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4833bbc-33", "ovs_interfaceid": "c4833bbc-333e-46c2-a140-5e8d493d3973", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "697d9636-5a3c-4b4e-8a86-3ea919b04757", "address": "fa:16:3e:73:76:c3", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap697d9636-5a", "ovs_interfaceid": "697d9636-5a3c-4b4e-8a86-3ea919b04757", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b42049a4-5eda-4603-950e-36103e3b012a", "address": "fa:16:3e:20:21:9e", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb42049a4-5e", "ovs_interfaceid": "b42049a4-5eda-4603-950e-36103e3b012a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.575598] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062067, 'name': Rename_Task, 'duration_secs': 0.215695} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.575888] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 974.576177] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7b994755-2649-48a5-aa86-f418236ddf84 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.585037] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 974.585037] env[61987]: value = "task-1062068" [ 974.585037] env[61987]: _type = "Task" [ 974.585037] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.591101] env[61987]: DEBUG oslo_concurrency.lockutils [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "473e0b73-6fcb-44c1-abf7-47d30c58af1b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.591348] env[61987]: DEBUG oslo_concurrency.lockutils [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "473e0b73-6fcb-44c1-abf7-47d30c58af1b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.598691] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062068, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.608151] env[61987]: DEBUG nova.compute.manager [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 974.793784] env[61987]: DEBUG oslo_concurrency.lockutils [req-5e5d5f53-4d53-4c1f-9964-60c9e2a1686a req-f761e1e3-f7a1-4405-8cd9-f9c04f870f48 service nova] Releasing lock "refresh_cache-4d15848d-5265-4897-bfa7-f3ad4c02ed83" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 974.893796] env[61987]: DEBUG oslo_vmware.api [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062066, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.049315] env[61987]: DEBUG oslo_concurrency.lockutils [req-27892a78-01ab-426a-ae7b-3435e38ff1bf req-aaa32319-76b4-4c03-921c-317674f4dde7 service nova] Releasing lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.074305] env[61987]: DEBUG nova.network.neutron [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 975.094632] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062068, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.099949] env[61987]: DEBUG nova.compute.manager [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 975.137301] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.137543] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.139099] env[61987]: INFO nova.compute.claims [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 975.280722] env[61987]: DEBUG nova.network.neutron [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Updating instance_info_cache with network_info: [{"id": "e237b976-f4a7-4dd6-8003-9b80e7f8e5fe", "address": "fa:16:3e:2e:1e:47", "network": {"id": "d5ed1e9e-9033-441e-bfcb-2ac643d7f03f", "bridge": "br-int", "label": "tempest-ServersTestJSON-8466993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e62741abfb4471bbc7e11ae7d407ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39ab9baf-90cd-4fe2-8d56-434f8210fc19", "external-id": "nsx-vlan-transportzone-713", "segmentation_id": 713, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape237b976-f4", "ovs_interfaceid": "e237b976-f4a7-4dd6-8003-9b80e7f8e5fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.394392] env[61987]: DEBUG oslo_vmware.api [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062066, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.532492] env[61987]: DEBUG nova.compute.manager [req-7b1fd3a2-4b14-4460-aef4-66d167d365cf req-210433e9-70be-4b82-9d52-62fca4277855 service nova] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Received event network-vif-plugged-e237b976-f4a7-4dd6-8003-9b80e7f8e5fe {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 975.532492] env[61987]: DEBUG oslo_concurrency.lockutils [req-7b1fd3a2-4b14-4460-aef4-66d167d365cf req-210433e9-70be-4b82-9d52-62fca4277855 service nova] Acquiring lock "6ec74ee9-a73b-4eac-97f2-abac3b1e3191-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.532723] env[61987]: DEBUG oslo_concurrency.lockutils [req-7b1fd3a2-4b14-4460-aef4-66d167d365cf req-210433e9-70be-4b82-9d52-62fca4277855 service nova] Lock "6ec74ee9-a73b-4eac-97f2-abac3b1e3191-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.532951] env[61987]: DEBUG oslo_concurrency.lockutils [req-7b1fd3a2-4b14-4460-aef4-66d167d365cf req-210433e9-70be-4b82-9d52-62fca4277855 service nova] Lock "6ec74ee9-a73b-4eac-97f2-abac3b1e3191-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.533160] env[61987]: DEBUG nova.compute.manager [req-7b1fd3a2-4b14-4460-aef4-66d167d365cf req-210433e9-70be-4b82-9d52-62fca4277855 service nova] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] No waiting events found dispatching network-vif-plugged-e237b976-f4a7-4dd6-8003-9b80e7f8e5fe {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 975.533344] env[61987]: WARNING nova.compute.manager [req-7b1fd3a2-4b14-4460-aef4-66d167d365cf req-210433e9-70be-4b82-9d52-62fca4277855 service nova] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Received unexpected event network-vif-plugged-e237b976-f4a7-4dd6-8003-9b80e7f8e5fe for instance with vm_state building and task_state spawning. [ 975.533511] env[61987]: DEBUG nova.compute.manager [req-7b1fd3a2-4b14-4460-aef4-66d167d365cf req-210433e9-70be-4b82-9d52-62fca4277855 service nova] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Received event network-changed-e237b976-f4a7-4dd6-8003-9b80e7f8e5fe {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 975.533700] env[61987]: DEBUG nova.compute.manager [req-7b1fd3a2-4b14-4460-aef4-66d167d365cf req-210433e9-70be-4b82-9d52-62fca4277855 service nova] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Refreshing instance network info cache due to event network-changed-e237b976-f4a7-4dd6-8003-9b80e7f8e5fe. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 975.533881] env[61987]: DEBUG oslo_concurrency.lockutils [req-7b1fd3a2-4b14-4460-aef4-66d167d365cf req-210433e9-70be-4b82-9d52-62fca4277855 service nova] Acquiring lock "refresh_cache-6ec74ee9-a73b-4eac-97f2-abac3b1e3191" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.593363] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062068, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.624773] env[61987]: DEBUG oslo_concurrency.lockutils [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.784243] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Releasing lock "refresh_cache-6ec74ee9-a73b-4eac-97f2-abac3b1e3191" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.784588] env[61987]: DEBUG nova.compute.manager [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Instance network_info: |[{"id": "e237b976-f4a7-4dd6-8003-9b80e7f8e5fe", "address": "fa:16:3e:2e:1e:47", "network": {"id": "d5ed1e9e-9033-441e-bfcb-2ac643d7f03f", "bridge": "br-int", "label": "tempest-ServersTestJSON-8466993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e62741abfb4471bbc7e11ae7d407ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39ab9baf-90cd-4fe2-8d56-434f8210fc19", "external-id": "nsx-vlan-transportzone-713", "segmentation_id": 713, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape237b976-f4", "ovs_interfaceid": "e237b976-f4a7-4dd6-8003-9b80e7f8e5fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 975.785080] env[61987]: DEBUG oslo_concurrency.lockutils [req-7b1fd3a2-4b14-4460-aef4-66d167d365cf req-210433e9-70be-4b82-9d52-62fca4277855 service nova] Acquired lock "refresh_cache-6ec74ee9-a73b-4eac-97f2-abac3b1e3191" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.785198] env[61987]: DEBUG nova.network.neutron [req-7b1fd3a2-4b14-4460-aef4-66d167d365cf req-210433e9-70be-4b82-9d52-62fca4277855 service nova] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Refreshing network info cache for port e237b976-f4a7-4dd6-8003-9b80e7f8e5fe {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 975.786911] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2e:1e:47', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '39ab9baf-90cd-4fe2-8d56-434f8210fc19', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e237b976-f4a7-4dd6-8003-9b80e7f8e5fe', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 975.794687] env[61987]: DEBUG oslo.service.loopingcall [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 975.795604] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 975.796332] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7e1d5abc-4e15-4f56-8a9e-5f7e45b51a7a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.818987] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 975.818987] env[61987]: value = "task-1062069" [ 975.818987] env[61987]: _type = "Task" [ 975.818987] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.831034] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062069, 'name': CreateVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.894158] env[61987]: DEBUG oslo_vmware.api [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062066, 'name': ReconfigVM_Task, 'duration_secs': 1.547266} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.894697] env[61987]: DEBUG oslo_concurrency.lockutils [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.895365] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Reconfigured VM to attach interface {{(pid=61987) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 976.093297] env[61987]: DEBUG oslo_vmware.api [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062068, 'name': PowerOnVM_Task, 'duration_secs': 1.477039} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.093579] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 976.093793] env[61987]: INFO nova.compute.manager [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Took 9.41 seconds to spawn the instance on the hypervisor. [ 976.093981] env[61987]: DEBUG nova.compute.manager [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 976.094837] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15fad3af-d82b-4399-9436-d4da1de5fdce {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.312226] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff263888-1e74-42e5-bc3d-c13f08f3979e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.320208] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffcf20a7-c505-481d-a37d-326cdb4e9ce8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.331457] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062069, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.360580] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83527fc1-aad0-47b5-a680-0a234d7e03aa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.368209] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b6fa57-be94-410a-924f-ad31db57e85c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.389806] env[61987]: DEBUG nova.compute.provider_tree [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.402905] env[61987]: DEBUG oslo_concurrency.lockutils [None req-424b0231-8fb0-441e-87c7-0c7940b68e8e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "interface-efdf62cc-9a3f-4b6b-a119-29a8096848ee-b42049a4-5eda-4603-950e-36103e3b012a" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.781s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.603090] env[61987]: DEBUG nova.network.neutron [req-7b1fd3a2-4b14-4460-aef4-66d167d365cf req-210433e9-70be-4b82-9d52-62fca4277855 service nova] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Updated VIF entry in instance network info cache for port e237b976-f4a7-4dd6-8003-9b80e7f8e5fe. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 976.603389] env[61987]: DEBUG nova.network.neutron [req-7b1fd3a2-4b14-4460-aef4-66d167d365cf req-210433e9-70be-4b82-9d52-62fca4277855 service nova] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Updating instance_info_cache with network_info: [{"id": "e237b976-f4a7-4dd6-8003-9b80e7f8e5fe", "address": "fa:16:3e:2e:1e:47", "network": {"id": "d5ed1e9e-9033-441e-bfcb-2ac643d7f03f", "bridge": "br-int", "label": "tempest-ServersTestJSON-8466993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e62741abfb4471bbc7e11ae7d407ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39ab9baf-90cd-4fe2-8d56-434f8210fc19", "external-id": "nsx-vlan-transportzone-713", "segmentation_id": 713, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape237b976-f4", "ovs_interfaceid": "e237b976-f4a7-4dd6-8003-9b80e7f8e5fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.610564] env[61987]: INFO nova.compute.manager [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Took 19.16 seconds to build instance. [ 976.832692] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062069, 'name': CreateVM_Task, 'duration_secs': 0.605196} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.832890] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 976.833540] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.833722] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.834068] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 976.834324] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d152bae-28fd-46aa-80f0-0da4e8bfacf5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.838726] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 976.838726] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52bdd89d-e411-e882-ac1f-e2c5a53e4d3c" [ 976.838726] env[61987]: _type = "Task" [ 976.838726] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.846831] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52bdd89d-e411-e882-ac1f-e2c5a53e4d3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.892586] env[61987]: DEBUG nova.scheduler.client.report [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 977.135434] env[61987]: DEBUG oslo_concurrency.lockutils [req-7b1fd3a2-4b14-4460-aef4-66d167d365cf req-210433e9-70be-4b82-9d52-62fca4277855 service nova] Releasing lock "refresh_cache-6ec74ee9-a73b-4eac-97f2-abac3b1e3191" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.135434] env[61987]: DEBUG oslo_concurrency.lockutils [None req-99cbc86a-d5c2-4961-b225-aa23eb2c0cd7 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "4dc624ca-c23c-4c4f-8481-45f954e1c068" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.663s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.349772] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52bdd89d-e411-e882-ac1f-e2c5a53e4d3c, 'name': SearchDatastore_Task, 'duration_secs': 0.022087} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.350093] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.350345] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 977.350596] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.350751] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.350937] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 977.351223] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4f485340-b1cb-4f15-a0fb-dc9d91a635f4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.366154] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 977.366623] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 977.367610] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab8d93b9-5246-4002-a0c6-ee25555a324b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.373731] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 977.373731] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]528b88b1-86e0-3776-ed9a-f8a7ce715c56" [ 977.373731] env[61987]: _type = "Task" [ 977.373731] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.383029] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]528b88b1-86e0-3776-ed9a-f8a7ce715c56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.398399] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.261s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.399198] env[61987]: DEBUG nova.compute.manager [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 977.403386] env[61987]: DEBUG oslo_concurrency.lockutils [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.779s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.404886] env[61987]: INFO nova.compute.claims [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 977.568822] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.569091] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.886085] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]528b88b1-86e0-3776-ed9a-f8a7ce715c56, 'name': SearchDatastore_Task, 'duration_secs': 0.023013} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.886377] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b47b41e-bc24-4fe7-b755-8aa380afb244 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.891378] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 977.891378] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52bfe004-90bb-f15a-e7c6-3d810faea344" [ 977.891378] env[61987]: _type = "Task" [ 977.891378] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.899031] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52bfe004-90bb-f15a-e7c6-3d810faea344, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.909550] env[61987]: DEBUG nova.compute.utils [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 977.911150] env[61987]: DEBUG nova.compute.manager [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 977.911425] env[61987]: DEBUG nova.network.neutron [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 977.956947] env[61987]: DEBUG nova.policy [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '914fdab15c68437db171f9710ea8ae85', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3a234a62da0e498fbe613fbcaaec3201', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 978.007248] env[61987]: DEBUG oslo_concurrency.lockutils [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "interface-efdf62cc-9a3f-4b6b-a119-29a8096848ee-697d9636-5a3c-4b4e-8a86-3ea919b04757" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.007626] env[61987]: DEBUG oslo_concurrency.lockutils [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "interface-efdf62cc-9a3f-4b6b-a119-29a8096848ee-697d9636-5a3c-4b4e-8a86-3ea919b04757" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.077413] env[61987]: DEBUG nova.compute.utils [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 978.219699] env[61987]: DEBUG nova.network.neutron [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Successfully created port: 70bf0473-599e-4d1e-be30-efec5a322e37 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 978.403836] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52bfe004-90bb-f15a-e7c6-3d810faea344, 'name': SearchDatastore_Task, 'duration_secs': 0.022908} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.404246] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.404616] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 6ec74ee9-a73b-4eac-97f2-abac3b1e3191/6ec74ee9-a73b-4eac-97f2-abac3b1e3191.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 978.404987] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9d7ac7c2-dc5b-4eba-a5d5-d4cf377a7d1c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.413743] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 978.413743] env[61987]: value = "task-1062070" [ 978.413743] env[61987]: _type = "Task" [ 978.413743] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.418913] env[61987]: DEBUG nova.compute.manager [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 978.432222] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062070, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.515806] env[61987]: DEBUG oslo_concurrency.lockutils [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.515806] env[61987]: DEBUG oslo_concurrency.lockutils [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.515806] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b21f95-8220-4ef4-b458-c566b2180e17 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.538397] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-580223c5-f5d5-4ab7-bd36-fff80d5dc0e1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.577850] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Reconfiguring VM to detach interface {{(pid=61987) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 978.583020] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-46c4339e-4e58-4db1-a6e2-0993ce462a1d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.600320] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.031s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.601628] env[61987]: DEBUG nova.compute.manager [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Stashing vm_state: active {{(pid=61987) _prep_resize /opt/stack/nova/nova/compute/manager.py:6090}} [ 978.612365] env[61987]: DEBUG oslo_vmware.api [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 978.612365] env[61987]: value = "task-1062071" [ 978.612365] env[61987]: _type = "Task" [ 978.612365] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.621996] env[61987]: DEBUG oslo_vmware.api [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062071, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.696299] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a990e17f-59f9-4910-9f6b-fb193b8630c1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.705803] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f50e4226-9421-488a-a1fb-91893c2d2a95 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.738721] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d01bfe-37bb-4bc8-bee6-d8910a9d30f0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.747475] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6720eeb-c6c4-4723-84e3-81174bc0a661 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.763092] env[61987]: DEBUG nova.compute.provider_tree [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.925180] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062070, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.125913] env[61987]: DEBUG oslo_vmware.api [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062071, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.128035] env[61987]: DEBUG oslo_concurrency.lockutils [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.266305] env[61987]: DEBUG nova.scheduler.client.report [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 979.425039] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062070, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.672413} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.425694] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 6ec74ee9-a73b-4eac-97f2-abac3b1e3191/6ec74ee9-a73b-4eac-97f2-abac3b1e3191.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 979.425926] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 979.426201] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4a57bc5e-e2ee-4cf5-bb58-2b525e14150b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.431317] env[61987]: DEBUG nova.compute.manager [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 979.435115] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 979.435115] env[61987]: value = "task-1062072" [ 979.435115] env[61987]: _type = "Task" [ 979.435115] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.441943] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062072, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.459884] env[61987]: DEBUG nova.virt.hardware [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 979.460211] env[61987]: DEBUG nova.virt.hardware [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 979.460439] env[61987]: DEBUG nova.virt.hardware [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 979.460703] env[61987]: DEBUG nova.virt.hardware [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 979.460913] env[61987]: DEBUG nova.virt.hardware [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 979.461259] env[61987]: DEBUG nova.virt.hardware [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 979.461564] env[61987]: DEBUG nova.virt.hardware [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 979.461786] env[61987]: DEBUG nova.virt.hardware [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 979.462029] env[61987]: DEBUG nova.virt.hardware [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 979.462251] env[61987]: DEBUG nova.virt.hardware [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 979.462501] env[61987]: DEBUG nova.virt.hardware [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 979.463503] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd20b041-d668-4ad3-a5f4-4edbfa728816 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.471756] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fba7eb2-3fad-4361-ae19-38f1b888383e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.624256] env[61987]: DEBUG oslo_vmware.api [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062071, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.660071] env[61987]: DEBUG nova.compute.manager [req-05fef23c-ab37-4aba-92bc-ea888ea14739 req-b4af93f6-286a-4c3d-88b6-719c707506f7 service nova] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Received event network-vif-plugged-70bf0473-599e-4d1e-be30-efec5a322e37 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 979.660313] env[61987]: DEBUG oslo_concurrency.lockutils [req-05fef23c-ab37-4aba-92bc-ea888ea14739 req-b4af93f6-286a-4c3d-88b6-719c707506f7 service nova] Acquiring lock "aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.660603] env[61987]: DEBUG oslo_concurrency.lockutils [req-05fef23c-ab37-4aba-92bc-ea888ea14739 req-b4af93f6-286a-4c3d-88b6-719c707506f7 service nova] Lock "aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.660826] env[61987]: DEBUG oslo_concurrency.lockutils [req-05fef23c-ab37-4aba-92bc-ea888ea14739 req-b4af93f6-286a-4c3d-88b6-719c707506f7 service nova] Lock "aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.660924] env[61987]: DEBUG nova.compute.manager [req-05fef23c-ab37-4aba-92bc-ea888ea14739 req-b4af93f6-286a-4c3d-88b6-719c707506f7 service nova] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] No waiting events found dispatching network-vif-plugged-70bf0473-599e-4d1e-be30-efec5a322e37 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 979.661277] env[61987]: WARNING nova.compute.manager [req-05fef23c-ab37-4aba-92bc-ea888ea14739 req-b4af93f6-286a-4c3d-88b6-719c707506f7 service nova] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Received unexpected event network-vif-plugged-70bf0473-599e-4d1e-be30-efec5a322e37 for instance with vm_state building and task_state spawning. [ 979.667885] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.668121] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.668724] env[61987]: INFO nova.compute.manager [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Attaching volume fdf4dca1-67ae-4a18-911d-6151e71f109b to /dev/sdb [ 979.697741] env[61987]: DEBUG oslo_concurrency.lockutils [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquiring lock "e6236bf1-4068-4a6e-9787-cd812381e122" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.697741] env[61987]: DEBUG oslo_concurrency.lockutils [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lock "e6236bf1-4068-4a6e-9787-cd812381e122" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.708701] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49b62f52-fca4-4f6e-9383-f90a629884a1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.716440] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb656ae7-f6b4-45d4-9ae8-19c987f73ec7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.729971] env[61987]: DEBUG nova.virt.block_device [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Updating existing volume attachment record: 3cdcd3ff-de73-48ea-83dc-b8d9408e37cc {{(pid=61987) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 979.748943] env[61987]: DEBUG nova.network.neutron [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Successfully updated port: 70bf0473-599e-4d1e-be30-efec5a322e37 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 979.771612] env[61987]: DEBUG oslo_concurrency.lockutils [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.368s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.771967] env[61987]: DEBUG nova.compute.manager [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 979.774973] env[61987]: DEBUG oslo_concurrency.lockutils [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.648s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.944846] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062072, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068137} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.945149] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 979.945963] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d7997a-1d25-4a2b-be10-dde157e9385a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.967833] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 6ec74ee9-a73b-4eac-97f2-abac3b1e3191/6ec74ee9-a73b-4eac-97f2-abac3b1e3191.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 979.968155] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec9d02a9-d63c-43b3-9677-fa943b576069 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.986885] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 979.986885] env[61987]: value = "task-1062074" [ 979.986885] env[61987]: _type = "Task" [ 979.986885] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.994676] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062074, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.124148] env[61987]: DEBUG oslo_vmware.api [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062071, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.199876] env[61987]: DEBUG nova.compute.manager [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 980.255135] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "refresh_cache-aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.255331] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquired lock "refresh_cache-aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.255490] env[61987]: DEBUG nova.network.neutron [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 980.276884] env[61987]: DEBUG nova.compute.utils [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 980.277689] env[61987]: DEBUG nova.compute.manager [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 980.277855] env[61987]: DEBUG nova.network.neutron [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 980.285603] env[61987]: INFO nova.compute.claims [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 980.340031] env[61987]: DEBUG nova.policy [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '914fdab15c68437db171f9710ea8ae85', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3a234a62da0e498fbe613fbcaaec3201', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 980.497718] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062074, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.628129] env[61987]: DEBUG oslo_vmware.api [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062071, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.639086] env[61987]: DEBUG nova.network.neutron [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Successfully created port: 5e6a927f-3d25-42bc-86b7-39767ddfbd98 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 980.725888] env[61987]: DEBUG oslo_concurrency.lockutils [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.783650] env[61987]: DEBUG nova.compute.manager [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 980.790667] env[61987]: DEBUG nova.network.neutron [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 980.793931] env[61987]: INFO nova.compute.resource_tracker [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Updating resource usage from migration a799cdd4-fc8b-40af-a2f0-b591c5463c7d [ 980.937268] env[61987]: DEBUG nova.network.neutron [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Updating instance_info_cache with network_info: [{"id": "70bf0473-599e-4d1e-be30-efec5a322e37", "address": "fa:16:3e:5e:ec:e5", "network": {"id": "4ed4fbdc-05b0-44f8-a3d5-b7b288abd5e0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2067313775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a234a62da0e498fbe613fbcaaec3201", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70bf0473-59", "ovs_interfaceid": "70bf0473-599e-4d1e-be30-efec5a322e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.958748] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7376fa1c-eb9d-48ff-be49-67daae0b25b6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.966749] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30d45e47-9e52-47f3-9d56-dd58cf7c4a45 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.001921] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4565452-f2eb-4a0d-bdfd-ee7624eada3d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.012567] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c73f6aa1-ab4a-4d3f-b20b-85b262ac1b83 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.016767] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062074, 'name': ReconfigVM_Task, 'duration_secs': 0.815498} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.017192] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 6ec74ee9-a73b-4eac-97f2-abac3b1e3191/6ec74ee9-a73b-4eac-97f2-abac3b1e3191.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 981.018143] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-448990f0-bacf-4e31-9e14-ccf5dbf5c7f8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.027625] env[61987]: DEBUG nova.compute.provider_tree [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 981.029707] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 981.029707] env[61987]: value = "task-1062075" [ 981.029707] env[61987]: _type = "Task" [ 981.029707] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.037747] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062075, 'name': Rename_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.126226] env[61987]: DEBUG oslo_vmware.api [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062071, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.439852] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Releasing lock "refresh_cache-aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.439988] env[61987]: DEBUG nova.compute.manager [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Instance network_info: |[{"id": "70bf0473-599e-4d1e-be30-efec5a322e37", "address": "fa:16:3e:5e:ec:e5", "network": {"id": "4ed4fbdc-05b0-44f8-a3d5-b7b288abd5e0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2067313775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a234a62da0e498fbe613fbcaaec3201", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70bf0473-59", "ovs_interfaceid": "70bf0473-599e-4d1e-be30-efec5a322e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 981.440570] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5e:ec:e5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ddfb706a-add1-4e16-9ac4-d20b16a1df6d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '70bf0473-599e-4d1e-be30-efec5a322e37', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 981.447918] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Creating folder: Project (3a234a62da0e498fbe613fbcaaec3201). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 981.448241] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c357c2ab-d5eb-4702-b6be-c36280854c1c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.459525] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Created folder: Project (3a234a62da0e498fbe613fbcaaec3201) in parent group-v234219. [ 981.459686] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Creating folder: Instances. Parent ref: group-v234404. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 981.459922] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a8dc6ca7-adb1-44b7-9e16-bd9c15130afe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.468503] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Created folder: Instances in parent group-v234404. [ 981.468765] env[61987]: DEBUG oslo.service.loopingcall [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.469062] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 981.469283] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d3a8ae74-f627-467b-ad0e-0e913dff3d17 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.487980] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 981.487980] env[61987]: value = "task-1062078" [ 981.487980] env[61987]: _type = "Task" [ 981.487980] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.495347] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062078, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.531951] env[61987]: DEBUG nova.scheduler.client.report [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 981.544383] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062075, 'name': Rename_Task, 'duration_secs': 0.168216} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.544673] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 981.545012] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-db24aee2-451d-4b29-8cf2-fedc2aff9b3c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.551305] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 981.551305] env[61987]: value = "task-1062079" [ 981.551305] env[61987]: _type = "Task" [ 981.551305] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.559328] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062079, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.629228] env[61987]: DEBUG oslo_vmware.api [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062071, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.686173] env[61987]: DEBUG nova.compute.manager [req-7f318409-4aa7-4fdc-9f79-82bd3a804ce6 req-e83caaed-e6b4-471b-a0bd-19c39f7a4617 service nova] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Received event network-changed-70bf0473-599e-4d1e-be30-efec5a322e37 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 981.686424] env[61987]: DEBUG nova.compute.manager [req-7f318409-4aa7-4fdc-9f79-82bd3a804ce6 req-e83caaed-e6b4-471b-a0bd-19c39f7a4617 service nova] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Refreshing instance network info cache due to event network-changed-70bf0473-599e-4d1e-be30-efec5a322e37. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 981.686686] env[61987]: DEBUG oslo_concurrency.lockutils [req-7f318409-4aa7-4fdc-9f79-82bd3a804ce6 req-e83caaed-e6b4-471b-a0bd-19c39f7a4617 service nova] Acquiring lock "refresh_cache-aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.686851] env[61987]: DEBUG oslo_concurrency.lockutils [req-7f318409-4aa7-4fdc-9f79-82bd3a804ce6 req-e83caaed-e6b4-471b-a0bd-19c39f7a4617 service nova] Acquired lock "refresh_cache-aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.687034] env[61987]: DEBUG nova.network.neutron [req-7f318409-4aa7-4fdc-9f79-82bd3a804ce6 req-e83caaed-e6b4-471b-a0bd-19c39f7a4617 service nova] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Refreshing network info cache for port 70bf0473-599e-4d1e-be30-efec5a322e37 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 981.795382] env[61987]: DEBUG nova.compute.manager [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 981.822369] env[61987]: DEBUG nova.virt.hardware [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 981.822735] env[61987]: DEBUG nova.virt.hardware [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 981.822932] env[61987]: DEBUG nova.virt.hardware [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 981.823143] env[61987]: DEBUG nova.virt.hardware [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 981.823302] env[61987]: DEBUG nova.virt.hardware [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 981.823455] env[61987]: DEBUG nova.virt.hardware [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 981.823663] env[61987]: DEBUG nova.virt.hardware [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 981.823829] env[61987]: DEBUG nova.virt.hardware [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 981.824009] env[61987]: DEBUG nova.virt.hardware [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 981.824183] env[61987]: DEBUG nova.virt.hardware [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 981.824363] env[61987]: DEBUG nova.virt.hardware [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 981.825282] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4165b959-2800-40cd-8d3d-a5f7ff65be88 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.833076] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4101bcd6-2801-476d-946f-8382844aefb3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.998336] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062078, 'name': CreateVM_Task, 'duration_secs': 0.318358} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.998708] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 981.999273] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.999466] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.999822] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 982.000102] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da4fb9e2-c322-4cf4-aa37-3c25e0812bc0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.004838] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 982.004838] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5242c684-8573-0887-0631-8b7d702a9c8b" [ 982.004838] env[61987]: _type = "Task" [ 982.004838] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.012599] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5242c684-8573-0887-0631-8b7d702a9c8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.039687] env[61987]: DEBUG oslo_concurrency.lockutils [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.265s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.039975] env[61987]: INFO nova.compute.manager [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Migrating [ 982.049470] env[61987]: DEBUG oslo_concurrency.lockutils [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.324s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.051075] env[61987]: INFO nova.compute.claims [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 982.054506] env[61987]: DEBUG nova.compute.manager [req-b45c00a5-1adf-4206-9e59-bbc4bfb8ba95 req-4e9fb4a1-e814-4510-baa5-3955f7443e24 service nova] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Received event network-vif-plugged-5e6a927f-3d25-42bc-86b7-39767ddfbd98 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 982.054804] env[61987]: DEBUG oslo_concurrency.lockutils [req-b45c00a5-1adf-4206-9e59-bbc4bfb8ba95 req-4e9fb4a1-e814-4510-baa5-3955f7443e24 service nova] Acquiring lock "473e0b73-6fcb-44c1-abf7-47d30c58af1b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.055089] env[61987]: DEBUG oslo_concurrency.lockutils [req-b45c00a5-1adf-4206-9e59-bbc4bfb8ba95 req-4e9fb4a1-e814-4510-baa5-3955f7443e24 service nova] Lock "473e0b73-6fcb-44c1-abf7-47d30c58af1b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.055322] env[61987]: DEBUG oslo_concurrency.lockutils [req-b45c00a5-1adf-4206-9e59-bbc4bfb8ba95 req-4e9fb4a1-e814-4510-baa5-3955f7443e24 service nova] Lock "473e0b73-6fcb-44c1-abf7-47d30c58af1b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.055564] env[61987]: DEBUG nova.compute.manager [req-b45c00a5-1adf-4206-9e59-bbc4bfb8ba95 req-4e9fb4a1-e814-4510-baa5-3955f7443e24 service nova] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] No waiting events found dispatching network-vif-plugged-5e6a927f-3d25-42bc-86b7-39767ddfbd98 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 982.055801] env[61987]: WARNING nova.compute.manager [req-b45c00a5-1adf-4206-9e59-bbc4bfb8ba95 req-4e9fb4a1-e814-4510-baa5-3955f7443e24 service nova] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Received unexpected event network-vif-plugged-5e6a927f-3d25-42bc-86b7-39767ddfbd98 for instance with vm_state building and task_state spawning. [ 982.073742] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062079, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.126791] env[61987]: DEBUG oslo_vmware.api [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062071, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.389397] env[61987]: DEBUG nova.network.neutron [req-7f318409-4aa7-4fdc-9f79-82bd3a804ce6 req-e83caaed-e6b4-471b-a0bd-19c39f7a4617 service nova] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Updated VIF entry in instance network info cache for port 70bf0473-599e-4d1e-be30-efec5a322e37. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 982.389807] env[61987]: DEBUG nova.network.neutron [req-7f318409-4aa7-4fdc-9f79-82bd3a804ce6 req-e83caaed-e6b4-471b-a0bd-19c39f7a4617 service nova] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Updating instance_info_cache with network_info: [{"id": "70bf0473-599e-4d1e-be30-efec5a322e37", "address": "fa:16:3e:5e:ec:e5", "network": {"id": "4ed4fbdc-05b0-44f8-a3d5-b7b288abd5e0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2067313775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a234a62da0e498fbe613fbcaaec3201", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70bf0473-59", "ovs_interfaceid": "70bf0473-599e-4d1e-be30-efec5a322e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.517560] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5242c684-8573-0887-0631-8b7d702a9c8b, 'name': SearchDatastore_Task, 'duration_secs': 0.00831} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.517871] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.518134] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 982.518378] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.518534] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.518718] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 982.519013] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4e559dea-ad37-4deb-b467-ea10dbbbca5b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.526782] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 982.526965] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 982.527861] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e997c01-55a0-45d9-b608-9d4cd0fde75e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.533122] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 982.533122] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]524703af-b4fa-87b4-571c-e9224934b8fd" [ 982.533122] env[61987]: _type = "Task" [ 982.533122] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.541020] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524703af-b4fa-87b4-571c-e9224934b8fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.562775] env[61987]: DEBUG oslo_concurrency.lockutils [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "refresh_cache-4dc624ca-c23c-4c4f-8481-45f954e1c068" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.562974] env[61987]: DEBUG oslo_concurrency.lockutils [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired lock "refresh_cache-4dc624ca-c23c-4c4f-8481-45f954e1c068" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.563131] env[61987]: DEBUG nova.network.neutron [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 982.572242] env[61987]: DEBUG oslo_vmware.api [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062079, 'name': PowerOnVM_Task, 'duration_secs': 0.61634} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.573046] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 982.573267] env[61987]: INFO nova.compute.manager [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Took 9.14 seconds to spawn the instance on the hypervisor. [ 982.573455] env[61987]: DEBUG nova.compute.manager [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 982.574614] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8474586-b0b6-4425-ac73-27942cf0f821 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.614189] env[61987]: DEBUG nova.network.neutron [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Successfully updated port: 5e6a927f-3d25-42bc-86b7-39767ddfbd98 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 982.631430] env[61987]: DEBUG oslo_vmware.api [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062071, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.807107] env[61987]: DEBUG nova.objects.instance [None req-e439a139-062f-4e47-9b66-93f430ce7a41 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Lazy-loading 'flavor' on Instance uuid cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.892577] env[61987]: DEBUG oslo_concurrency.lockutils [req-7f318409-4aa7-4fdc-9f79-82bd3a804ce6 req-e83caaed-e6b4-471b-a0bd-19c39f7a4617 service nova] Releasing lock "refresh_cache-aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.043768] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524703af-b4fa-87b4-571c-e9224934b8fd, 'name': SearchDatastore_Task, 'duration_secs': 0.011789} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.044538] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-740e4691-0ab5-475c-95d7-10561842eb9a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.049263] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 983.049263] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52eab917-3c27-843b-314a-d2b8b2d97f69" [ 983.049263] env[61987]: _type = "Task" [ 983.049263] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.056599] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52eab917-3c27-843b-314a-d2b8b2d97f69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.093392] env[61987]: INFO nova.compute.manager [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Took 13.95 seconds to build instance. [ 983.118966] env[61987]: DEBUG oslo_concurrency.lockutils [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "refresh_cache-473e0b73-6fcb-44c1-abf7-47d30c58af1b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.118966] env[61987]: DEBUG oslo_concurrency.lockutils [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquired lock "refresh_cache-473e0b73-6fcb-44c1-abf7-47d30c58af1b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.118966] env[61987]: DEBUG nova.network.neutron [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 983.132138] env[61987]: DEBUG oslo_vmware.api [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062071, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.231109] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c6df3e-66ee-4a5a-9887-b551af49192c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.241035] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d168577-fa6d-41a7-8529-0cab4cad164c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.272526] env[61987]: DEBUG nova.network.neutron [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Updating instance_info_cache with network_info: [{"id": "1735c272-060e-4a5d-836a-af4e2eed480a", "address": "fa:16:3e:69:d8:56", "network": {"id": "ca68bf96-3f18-45ee-80fb-64a1260ed996", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-733595377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd47eb44bc334bf3ae5813905903ecbc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a27fd90b-16a5-43af-bede-ae36762ece00", "external-id": "nsx-vlan-transportzone-197", "segmentation_id": 197, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1735c272-06", "ovs_interfaceid": "1735c272-060e-4a5d-836a-af4e2eed480a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.274138] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0918055-857c-4a20-bf94-d424c8842dc1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.281970] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40ee8e9d-6f90-4d0d-9c20-46492020ccea {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.296067] env[61987]: DEBUG nova.compute.provider_tree [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 983.311361] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e439a139-062f-4e47-9b66-93f430ce7a41 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Acquiring lock "refresh_cache-cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.311429] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e439a139-062f-4e47-9b66-93f430ce7a41 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Acquired lock "refresh_cache-cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.489459] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 983.489693] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 983.559816] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52eab917-3c27-843b-314a-d2b8b2d97f69, 'name': SearchDatastore_Task, 'duration_secs': 0.008656} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.560110] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.560424] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8/aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 983.560695] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-55a90d9d-4173-4c6f-aff6-0934e7b3b69f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.567432] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 983.567432] env[61987]: value = "task-1062081" [ 983.567432] env[61987]: _type = "Task" [ 983.567432] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.575742] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062081, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.594922] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b04f4c9c-a4fd-492d-9772-3fffbfb88b2a tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "6ec74ee9-a73b-4eac-97f2-abac3b1e3191" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.463s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.637235] env[61987]: DEBUG oslo_vmware.api [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062071, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.660261] env[61987]: DEBUG nova.network.neutron [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 983.665361] env[61987]: DEBUG nova.network.neutron [None req-e439a139-062f-4e47-9b66-93f430ce7a41 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 983.744152] env[61987]: DEBUG nova.compute.manager [req-565bbb64-b63d-4479-91cb-1e494b1d521a req-4487dff1-37cd-490b-90f6-42e074af80cd service nova] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Received event network-changed-5e6a927f-3d25-42bc-86b7-39767ddfbd98 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 983.744152] env[61987]: DEBUG nova.compute.manager [req-565bbb64-b63d-4479-91cb-1e494b1d521a req-4487dff1-37cd-490b-90f6-42e074af80cd service nova] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Refreshing instance network info cache due to event network-changed-5e6a927f-3d25-42bc-86b7-39767ddfbd98. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 983.744936] env[61987]: DEBUG oslo_concurrency.lockutils [req-565bbb64-b63d-4479-91cb-1e494b1d521a req-4487dff1-37cd-490b-90f6-42e074af80cd service nova] Acquiring lock "refresh_cache-473e0b73-6fcb-44c1-abf7-47d30c58af1b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.777868] env[61987]: DEBUG oslo_concurrency.lockutils [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Releasing lock "refresh_cache-4dc624ca-c23c-4c4f-8481-45f954e1c068" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.799641] env[61987]: DEBUG nova.scheduler.client.report [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 983.871481] env[61987]: DEBUG nova.network.neutron [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Updating instance_info_cache with network_info: [{"id": "5e6a927f-3d25-42bc-86b7-39767ddfbd98", "address": "fa:16:3e:c8:53:17", "network": {"id": "4ed4fbdc-05b0-44f8-a3d5-b7b288abd5e0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2067313775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a234a62da0e498fbe613fbcaaec3201", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e6a927f-3d", "ovs_interfaceid": "5e6a927f-3d25-42bc-86b7-39767ddfbd98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.996319] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 983.996466] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Starting heal instance info cache {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10401}} [ 983.996567] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Rebuilding the list of instances to heal {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10405}} [ 984.077639] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062081, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.133482] env[61987]: DEBUG oslo_vmware.api [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062071, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.290814] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Volume attach. Driver type: vmdk {{(pid=61987) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 984.290814] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234403', 'volume_id': 'fdf4dca1-67ae-4a18-911d-6151e71f109b', 'name': 'volume-fdf4dca1-67ae-4a18-911d-6151e71f109b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ae1396b3-c28b-4d1c-9a3a-3dc2170847ca', 'attached_at': '', 'detached_at': '', 'volume_id': 'fdf4dca1-67ae-4a18-911d-6151e71f109b', 'serial': 'fdf4dca1-67ae-4a18-911d-6151e71f109b'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 984.292034] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e86a7e3-636e-44bd-b6e0-0d583c771b47 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.309525] env[61987]: DEBUG oslo_concurrency.lockutils [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.260s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.310092] env[61987]: DEBUG nova.compute.manager [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 984.314081] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c35905c-faed-4b87-a5f4-f75ea3510aac {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.342379] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] volume-fdf4dca1-67ae-4a18-911d-6151e71f109b/volume-fdf4dca1-67ae-4a18-911d-6151e71f109b.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 984.343459] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c6d3734-9713-4c99-a734-3e220db0d02a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.364778] env[61987]: DEBUG oslo_vmware.api [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 984.364778] env[61987]: value = "task-1062082" [ 984.364778] env[61987]: _type = "Task" [ 984.364778] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.373102] env[61987]: DEBUG oslo_concurrency.lockutils [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Releasing lock "refresh_cache-473e0b73-6fcb-44c1-abf7-47d30c58af1b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.373354] env[61987]: DEBUG nova.compute.manager [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Instance network_info: |[{"id": "5e6a927f-3d25-42bc-86b7-39767ddfbd98", "address": "fa:16:3e:c8:53:17", "network": {"id": "4ed4fbdc-05b0-44f8-a3d5-b7b288abd5e0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2067313775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a234a62da0e498fbe613fbcaaec3201", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e6a927f-3d", "ovs_interfaceid": "5e6a927f-3d25-42bc-86b7-39767ddfbd98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 984.374047] env[61987]: DEBUG oslo_vmware.api [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062082, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.374047] env[61987]: DEBUG oslo_concurrency.lockutils [req-565bbb64-b63d-4479-91cb-1e494b1d521a req-4487dff1-37cd-490b-90f6-42e074af80cd service nova] Acquired lock "refresh_cache-473e0b73-6fcb-44c1-abf7-47d30c58af1b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.374047] env[61987]: DEBUG nova.network.neutron [req-565bbb64-b63d-4479-91cb-1e494b1d521a req-4487dff1-37cd-490b-90f6-42e074af80cd service nova] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Refreshing network info cache for port 5e6a927f-3d25-42bc-86b7-39767ddfbd98 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 984.375136] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c8:53:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ddfb706a-add1-4e16-9ac4-d20b16a1df6d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5e6a927f-3d25-42bc-86b7-39767ddfbd98', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 984.382454] env[61987]: DEBUG oslo.service.loopingcall [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 984.385365] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 984.386167] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c5de0337-ddfa-4f63-862c-2a58db1233ad {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.409202] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 984.409202] env[61987]: value = "task-1062083" [ 984.409202] env[61987]: _type = "Task" [ 984.409202] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.417062] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062083, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.438471] env[61987]: DEBUG nova.network.neutron [None req-e439a139-062f-4e47-9b66-93f430ce7a41 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Updating instance_info_cache with network_info: [{"id": "feaf0c72-3782-4054-907b-f98cc72cd02a", "address": "fa:16:3e:13:f5:64", "network": {"id": "b45b790f-b724-4a9a-b54b-8f8319b13a88", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1186325730-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7f68eb92e124eab9beb7b60ddcc1df3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfeaf0c72-37", "ovs_interfaceid": "feaf0c72-3782-4054-907b-f98cc72cd02a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.504315] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Skipping network cache update for instance because it is Building. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10414}} [ 984.504488] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Skipping network cache update for instance because it is Building. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10414}} [ 984.504624] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Skipping network cache update for instance because it is Building. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10414}} [ 984.530887] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "refresh_cache-84a4a69b-04db-4546-ba89-9039e382a0c4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.531055] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquired lock "refresh_cache-84a4a69b-04db-4546-ba89-9039e382a0c4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.531208] env[61987]: DEBUG nova.network.neutron [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Forcefully refreshing network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 984.531361] env[61987]: DEBUG nova.objects.instance [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lazy-loading 'info_cache' on Instance uuid 84a4a69b-04db-4546-ba89-9039e382a0c4 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.577573] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062081, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.537763} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.577827] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8/aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 984.578110] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 984.578359] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cd913c49-4908-40a1-8370-f8e5b775149f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.585355] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 984.585355] env[61987]: value = "task-1062084" [ 984.585355] env[61987]: _type = "Task" [ 984.585355] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.588684] env[61987]: DEBUG nova.network.neutron [req-565bbb64-b63d-4479-91cb-1e494b1d521a req-4487dff1-37cd-490b-90f6-42e074af80cd service nova] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Updated VIF entry in instance network info cache for port 5e6a927f-3d25-42bc-86b7-39767ddfbd98. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 984.589016] env[61987]: DEBUG nova.network.neutron [req-565bbb64-b63d-4479-91cb-1e494b1d521a req-4487dff1-37cd-490b-90f6-42e074af80cd service nova] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Updating instance_info_cache with network_info: [{"id": "5e6a927f-3d25-42bc-86b7-39767ddfbd98", "address": "fa:16:3e:c8:53:17", "network": {"id": "4ed4fbdc-05b0-44f8-a3d5-b7b288abd5e0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2067313775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a234a62da0e498fbe613fbcaaec3201", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e6a927f-3d", "ovs_interfaceid": "5e6a927f-3d25-42bc-86b7-39767ddfbd98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.594833] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062084, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.635946] env[61987]: DEBUG oslo_vmware.api [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062071, 'name': ReconfigVM_Task, 'duration_secs': 5.743282} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.636266] env[61987]: DEBUG oslo_concurrency.lockutils [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.636501] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Reconfigured VM to detach interface {{(pid=61987) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 984.755078] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "6ec74ee9-a73b-4eac-97f2-abac3b1e3191" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.755460] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "6ec74ee9-a73b-4eac-97f2-abac3b1e3191" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.755756] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "6ec74ee9-a73b-4eac-97f2-abac3b1e3191-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.755997] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "6ec74ee9-a73b-4eac-97f2-abac3b1e3191-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.756231] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "6ec74ee9-a73b-4eac-97f2-abac3b1e3191-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.758761] env[61987]: INFO nova.compute.manager [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Terminating instance [ 984.814928] env[61987]: DEBUG nova.compute.utils [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 984.816948] env[61987]: DEBUG nova.compute.manager [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 984.817169] env[61987]: DEBUG nova.network.neutron [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 984.861800] env[61987]: DEBUG nova.policy [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06accdea3e014922af1f011175c36bc6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '630439422a354de0b32ac5eefaa990b9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 984.875661] env[61987]: DEBUG oslo_vmware.api [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062082, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.920501] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062083, 'name': CreateVM_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.941409] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e439a139-062f-4e47-9b66-93f430ce7a41 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Releasing lock "refresh_cache-cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.941670] env[61987]: DEBUG nova.compute.manager [None req-e439a139-062f-4e47-9b66-93f430ce7a41 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Inject network info {{(pid=61987) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7692}} [ 984.941998] env[61987]: DEBUG nova.compute.manager [None req-e439a139-062f-4e47-9b66-93f430ce7a41 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] network_info to inject: |[{"id": "feaf0c72-3782-4054-907b-f98cc72cd02a", "address": "fa:16:3e:13:f5:64", "network": {"id": "b45b790f-b724-4a9a-b54b-8f8319b13a88", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1186325730-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7f68eb92e124eab9beb7b60ddcc1df3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfeaf0c72-37", "ovs_interfaceid": "feaf0c72-3782-4054-907b-f98cc72cd02a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7693}} [ 984.947437] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e439a139-062f-4e47-9b66-93f430ce7a41 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Reconfiguring VM instance to set the machine id {{(pid=61987) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 984.947437] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-91c7aea6-636a-40d0-9a94-993f59a0b5a7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.963823] env[61987]: DEBUG oslo_vmware.api [None req-e439a139-062f-4e47-9b66-93f430ce7a41 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Waiting for the task: (returnval){ [ 984.963823] env[61987]: value = "task-1062085" [ 984.963823] env[61987]: _type = "Task" [ 984.963823] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.973999] env[61987]: DEBUG oslo_vmware.api [None req-e439a139-062f-4e47-9b66-93f430ce7a41 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062085, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.091846] env[61987]: DEBUG oslo_concurrency.lockutils [req-565bbb64-b63d-4479-91cb-1e494b1d521a req-4487dff1-37cd-490b-90f6-42e074af80cd service nova] Releasing lock "refresh_cache-473e0b73-6fcb-44c1-abf7-47d30c58af1b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.092237] env[61987]: DEBUG nova.compute.manager [req-565bbb64-b63d-4479-91cb-1e494b1d521a req-4487dff1-37cd-490b-90f6-42e074af80cd service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Received event network-changed-feaf0c72-3782-4054-907b-f98cc72cd02a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 985.092237] env[61987]: DEBUG nova.compute.manager [req-565bbb64-b63d-4479-91cb-1e494b1d521a req-4487dff1-37cd-490b-90f6-42e074af80cd service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Refreshing instance network info cache due to event network-changed-feaf0c72-3782-4054-907b-f98cc72cd02a. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 985.092424] env[61987]: DEBUG oslo_concurrency.lockutils [req-565bbb64-b63d-4479-91cb-1e494b1d521a req-4487dff1-37cd-490b-90f6-42e074af80cd service nova] Acquiring lock "refresh_cache-cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.092576] env[61987]: DEBUG oslo_concurrency.lockutils [req-565bbb64-b63d-4479-91cb-1e494b1d521a req-4487dff1-37cd-490b-90f6-42e074af80cd service nova] Acquired lock "refresh_cache-cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.092837] env[61987]: DEBUG nova.network.neutron [req-565bbb64-b63d-4479-91cb-1e494b1d521a req-4487dff1-37cd-490b-90f6-42e074af80cd service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Refreshing network info cache for port feaf0c72-3782-4054-907b-f98cc72cd02a {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 985.097560] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062084, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.506616} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.098291] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 985.098903] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e9dcf3e-549b-405a-8d60-23e8ce55fe27 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.121863] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8/aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 985.122527] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94f01ba0-d069-4b98-861c-bf7083b17595 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.142763] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 985.142763] env[61987]: value = "task-1062086" [ 985.142763] env[61987]: _type = "Task" [ 985.142763] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.153531] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062086, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.166973] env[61987]: DEBUG nova.objects.instance [None req-4750d3ba-3203-4f69-a7e3-317423f1424f tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Lazy-loading 'flavor' on Instance uuid cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.228743] env[61987]: DEBUG nova.network.neutron [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Successfully created port: e9aa9f0a-f5e5-4d71-bddd-c817a7b4baf1 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 985.263613] env[61987]: DEBUG nova.compute.manager [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 985.263868] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 985.264948] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f840f0e-cef4-4cd4-957b-d63eb45c7449 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.273734] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 985.274014] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2d3484dc-47ea-4b85-8a0e-c22a798fe4e1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.280108] env[61987]: DEBUG oslo_vmware.api [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 985.280108] env[61987]: value = "task-1062087" [ 985.280108] env[61987]: _type = "Task" [ 985.280108] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.289757] env[61987]: DEBUG oslo_vmware.api [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062087, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.304199] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6b9985e-bbd5-4d9c-8c9b-4647c28434c8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.323733] env[61987]: DEBUG nova.compute.manager [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 985.326552] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Updating instance '4dc624ca-c23c-4c4f-8481-45f954e1c068' progress to 0 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 985.375890] env[61987]: DEBUG oslo_vmware.api [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062082, 'name': ReconfigVM_Task, 'duration_secs': 0.632495} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.376250] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Reconfigured VM instance instance-0000005a to attach disk [datastore2] volume-fdf4dca1-67ae-4a18-911d-6151e71f109b/volume-fdf4dca1-67ae-4a18-911d-6151e71f109b.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 985.381236] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d40623d-0139-4e84-8ff9-f854477b83cf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.397418] env[61987]: DEBUG oslo_vmware.api [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 985.397418] env[61987]: value = "task-1062088" [ 985.397418] env[61987]: _type = "Task" [ 985.397418] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.407284] env[61987]: DEBUG oslo_vmware.api [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062088, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.421842] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062083, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.476118] env[61987]: DEBUG oslo_vmware.api [None req-e439a139-062f-4e47-9b66-93f430ce7a41 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062085, 'name': ReconfigVM_Task, 'duration_secs': 0.426514} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.476430] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e439a139-062f-4e47-9b66-93f430ce7a41 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Reconfigured VM instance to set the machine id {{(pid=61987) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 985.652777] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062086, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.671789] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4750d3ba-3203-4f69-a7e3-317423f1424f tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Acquiring lock "refresh_cache-cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.775607] env[61987]: DEBUG nova.compute.manager [req-3165e578-a394-4ef3-823e-416a0dc2870f req-6ce9fd45-2689-402e-8fa7-e3cf5466c2c8 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Received event network-vif-deleted-697d9636-5a3c-4b4e-8a86-3ea919b04757 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 985.775892] env[61987]: INFO nova.compute.manager [req-3165e578-a394-4ef3-823e-416a0dc2870f req-6ce9fd45-2689-402e-8fa7-e3cf5466c2c8 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Neutron deleted interface 697d9636-5a3c-4b4e-8a86-3ea919b04757; detaching it from the instance and deleting it from the info cache [ 985.776270] env[61987]: DEBUG nova.network.neutron [req-3165e578-a394-4ef3-823e-416a0dc2870f req-6ce9fd45-2689-402e-8fa7-e3cf5466c2c8 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Updating instance_info_cache with network_info: [{"id": "c4833bbc-333e-46c2-a140-5e8d493d3973", "address": "fa:16:3e:11:e5:35", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4833bbc-33", "ovs_interfaceid": "c4833bbc-333e-46c2-a140-5e8d493d3973", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b42049a4-5eda-4603-950e-36103e3b012a", "address": "fa:16:3e:20:21:9e", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb42049a4-5e", "ovs_interfaceid": "b42049a4-5eda-4603-950e-36103e3b012a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.794440] env[61987]: DEBUG oslo_vmware.api [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062087, 'name': PowerOffVM_Task, 'duration_secs': 0.481774} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.795491] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 985.795823] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 985.796479] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8b38ff45-3acc-43a1-80aa-e3940a10b5d5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.835199] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 985.835472] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-481a65f0-4836-41f7-afd5-328a041743ee {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.842652] env[61987]: DEBUG oslo_vmware.api [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 985.842652] env[61987]: value = "task-1062090" [ 985.842652] env[61987]: _type = "Task" [ 985.842652] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.854248] env[61987]: DEBUG oslo_vmware.api [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062090, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.857540] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 985.857749] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 985.857945] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Deleting the datastore file [datastore2] 6ec74ee9-a73b-4eac-97f2-abac3b1e3191 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 985.858499] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c62714fb-3542-43d7-82f5-961cf1ae3176 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.866952] env[61987]: DEBUG oslo_vmware.api [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 985.866952] env[61987]: value = "task-1062091" [ 985.866952] env[61987]: _type = "Task" [ 985.866952] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.876258] env[61987]: DEBUG oslo_vmware.api [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062091, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.893639] env[61987]: DEBUG nova.network.neutron [req-565bbb64-b63d-4479-91cb-1e494b1d521a req-4487dff1-37cd-490b-90f6-42e074af80cd service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Updated VIF entry in instance network info cache for port feaf0c72-3782-4054-907b-f98cc72cd02a. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 985.894079] env[61987]: DEBUG nova.network.neutron [req-565bbb64-b63d-4479-91cb-1e494b1d521a req-4487dff1-37cd-490b-90f6-42e074af80cd service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Updating instance_info_cache with network_info: [{"id": "feaf0c72-3782-4054-907b-f98cc72cd02a", "address": "fa:16:3e:13:f5:64", "network": {"id": "b45b790f-b724-4a9a-b54b-8f8319b13a88", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1186325730-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7f68eb92e124eab9beb7b60ddcc1df3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfeaf0c72-37", "ovs_interfaceid": "feaf0c72-3782-4054-907b-f98cc72cd02a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.908456] env[61987]: DEBUG oslo_vmware.api [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062088, 'name': ReconfigVM_Task, 'duration_secs': 0.178577} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.908894] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234403', 'volume_id': 'fdf4dca1-67ae-4a18-911d-6151e71f109b', 'name': 'volume-fdf4dca1-67ae-4a18-911d-6151e71f109b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ae1396b3-c28b-4d1c-9a3a-3dc2170847ca', 'attached_at': '', 'detached_at': '', 'volume_id': 'fdf4dca1-67ae-4a18-911d-6151e71f109b', 'serial': 'fdf4dca1-67ae-4a18-911d-6151e71f109b'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 985.920975] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062083, 'name': CreateVM_Task, 'duration_secs': 1.042143} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.921205] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 985.921901] env[61987]: DEBUG oslo_concurrency.lockutils [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.922088] env[61987]: DEBUG oslo_concurrency.lockutils [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.922423] env[61987]: DEBUG oslo_concurrency.lockutils [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 985.922700] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5eb088fd-3dea-44a9-82b1-6e622eb2c447 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.928935] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 985.928935] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52cfeda6-f940-148f-bb7f-78a1a29c27bd" [ 985.928935] env[61987]: _type = "Task" [ 985.928935] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.938541] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52cfeda6-f940-148f-bb7f-78a1a29c27bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.022019] env[61987]: DEBUG oslo_concurrency.lockutils [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.022291] env[61987]: DEBUG oslo_concurrency.lockutils [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.022500] env[61987]: DEBUG nova.network.neutron [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 986.065238] env[61987]: DEBUG nova.compute.manager [req-48cad022-0bc4-4d4d-a47c-d0a16e474ef4 req-a0414416-8458-4cb2-a3f9-3956f2c9677f service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Received event network-vif-deleted-b42049a4-5eda-4603-950e-36103e3b012a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 986.065238] env[61987]: INFO nova.compute.manager [req-48cad022-0bc4-4d4d-a47c-d0a16e474ef4 req-a0414416-8458-4cb2-a3f9-3956f2c9677f service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Neutron deleted interface b42049a4-5eda-4603-950e-36103e3b012a; detaching it from the instance and deleting it from the info cache [ 986.065238] env[61987]: DEBUG nova.network.neutron [req-48cad022-0bc4-4d4d-a47c-d0a16e474ef4 req-a0414416-8458-4cb2-a3f9-3956f2c9677f service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Updating instance_info_cache with network_info: [{"id": "c4833bbc-333e-46c2-a140-5e8d493d3973", "address": "fa:16:3e:11:e5:35", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4833bbc-33", "ovs_interfaceid": "c4833bbc-333e-46c2-a140-5e8d493d3973", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.159499] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062086, 'name': ReconfigVM_Task, 'duration_secs': 0.627256} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.159799] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Reconfigured VM instance instance-00000063 to attach disk [datastore2] aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8/aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 986.160490] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9a93cb3f-88c9-49f2-998f-aee062b086eb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.167109] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 986.167109] env[61987]: value = "task-1062092" [ 986.167109] env[61987]: _type = "Task" [ 986.167109] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.176194] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062092, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.278855] env[61987]: DEBUG oslo_concurrency.lockutils [req-3165e578-a394-4ef3-823e-416a0dc2870f req-6ce9fd45-2689-402e-8fa7-e3cf5466c2c8 service nova] Acquiring lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.279069] env[61987]: DEBUG oslo_concurrency.lockutils [req-3165e578-a394-4ef3-823e-416a0dc2870f req-6ce9fd45-2689-402e-8fa7-e3cf5466c2c8 service nova] Acquired lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.279944] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e1066d-2b08-45b9-9ac9-11e24c5041bf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.299577] env[61987]: DEBUG oslo_concurrency.lockutils [req-3165e578-a394-4ef3-823e-416a0dc2870f req-6ce9fd45-2689-402e-8fa7-e3cf5466c2c8 service nova] Releasing lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.299851] env[61987]: WARNING nova.compute.manager [req-3165e578-a394-4ef3-823e-416a0dc2870f req-6ce9fd45-2689-402e-8fa7-e3cf5466c2c8 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Detach interface failed, port_id=697d9636-5a3c-4b4e-8a86-3ea919b04757, reason: No device with interface-id 697d9636-5a3c-4b4e-8a86-3ea919b04757 exists on VM: nova.exception.NotFound: No device with interface-id 697d9636-5a3c-4b4e-8a86-3ea919b04757 exists on VM [ 986.300900] env[61987]: DEBUG nova.network.neutron [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Updating instance_info_cache with network_info: [{"id": "c9f4e2e0-50b8-4e72-a06d-8c2729b200ed", "address": "fa:16:3e:74:cd:56", "network": {"id": "d5ed1e9e-9033-441e-bfcb-2ac643d7f03f", "bridge": "br-int", "label": "tempest-ServersTestJSON-8466993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e62741abfb4471bbc7e11ae7d407ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39ab9baf-90cd-4fe2-8d56-434f8210fc19", "external-id": "nsx-vlan-transportzone-713", "segmentation_id": 713, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc9f4e2e0-50", "ovs_interfaceid": "c9f4e2e0-50b8-4e72-a06d-8c2729b200ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.339897] env[61987]: DEBUG nova.compute.manager [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 986.352098] env[61987]: DEBUG oslo_vmware.api [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062090, 'name': PowerOffVM_Task, 'duration_secs': 0.180913} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.352369] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 986.352559] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Updating instance '4dc624ca-c23c-4c4f-8481-45f954e1c068' progress to 17 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 986.366943] env[61987]: DEBUG nova.virt.hardware [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 986.367326] env[61987]: DEBUG nova.virt.hardware [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 986.367521] env[61987]: DEBUG nova.virt.hardware [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 986.367817] env[61987]: DEBUG nova.virt.hardware [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 986.368226] env[61987]: DEBUG nova.virt.hardware [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 986.368226] env[61987]: DEBUG nova.virt.hardware [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 986.368324] env[61987]: DEBUG nova.virt.hardware [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 986.368429] env[61987]: DEBUG nova.virt.hardware [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 986.368597] env[61987]: DEBUG nova.virt.hardware [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 986.368760] env[61987]: DEBUG nova.virt.hardware [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 986.368940] env[61987]: DEBUG nova.virt.hardware [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 986.369880] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd747f0-f256-441b-8a73-8903466200db {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.383770] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159f2ca6-40cf-4fb3-82af-966f7bbbd783 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.388023] env[61987]: DEBUG oslo_vmware.api [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062091, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.339677} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.388289] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 986.388486] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 986.388773] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 986.388956] env[61987]: INFO nova.compute.manager [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Took 1.13 seconds to destroy the instance on the hypervisor. [ 986.389193] env[61987]: DEBUG oslo.service.loopingcall [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 986.389790] env[61987]: DEBUG nova.compute.manager [-] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 986.389897] env[61987]: DEBUG nova.network.neutron [-] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 986.399103] env[61987]: DEBUG oslo_concurrency.lockutils [req-565bbb64-b63d-4479-91cb-1e494b1d521a req-4487dff1-37cd-490b-90f6-42e074af80cd service nova] Releasing lock "refresh_cache-cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.400622] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4750d3ba-3203-4f69-a7e3-317423f1424f tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Acquired lock "refresh_cache-cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.439178] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52cfeda6-f940-148f-bb7f-78a1a29c27bd, 'name': SearchDatastore_Task, 'duration_secs': 0.016466} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.439512] env[61987]: DEBUG oslo_concurrency.lockutils [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.439767] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 986.440046] env[61987]: DEBUG oslo_concurrency.lockutils [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.440234] env[61987]: DEBUG oslo_concurrency.lockutils [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.440424] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 986.440690] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4dd35cff-110b-41a1-8c86-c111b7b3a23a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.448837] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 986.449044] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 986.451532] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6493b1aa-dce2-4f76-98cb-dd84f0d07cfc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.456563] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 986.456563] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]528ff4b6-67f9-1779-57a3-07f1cde374bb" [ 986.456563] env[61987]: _type = "Task" [ 986.456563] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.464142] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]528ff4b6-67f9-1779-57a3-07f1cde374bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.547961] env[61987]: DEBUG oslo_concurrency.lockutils [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.547961] env[61987]: DEBUG oslo_concurrency.lockutils [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.547961] env[61987]: DEBUG oslo_concurrency.lockutils [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.547961] env[61987]: DEBUG oslo_concurrency.lockutils [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.547961] env[61987]: DEBUG oslo_concurrency.lockutils [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.549196] env[61987]: INFO nova.compute.manager [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Terminating instance [ 986.568138] env[61987]: DEBUG oslo_concurrency.lockutils [req-48cad022-0bc4-4d4d-a47c-d0a16e474ef4 req-a0414416-8458-4cb2-a3f9-3956f2c9677f service nova] Acquiring lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.677125] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062092, 'name': Rename_Task, 'duration_secs': 0.374992} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.677425] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 986.677683] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bb02b7ff-e7a2-4aa8-981c-aec4e9fbc185 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.684596] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 986.684596] env[61987]: value = "task-1062093" [ 986.684596] env[61987]: _type = "Task" [ 986.684596] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.694333] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062093, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.803349] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Releasing lock "refresh_cache-84a4a69b-04db-4546-ba89-9039e382a0c4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.803575] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Updated the network info_cache for instance {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10472}} [ 986.803788] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 986.803959] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 986.804124] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 986.804282] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 986.804431] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 986.804612] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 986.804704] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61987) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11020}} [ 986.804877] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 986.862168] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 986.862475] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 986.862681] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 986.863062] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 986.863243] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 986.863407] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 986.863622] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 986.863792] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 986.863962] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 986.864239] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 986.864513] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 986.869695] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-120b5c4f-f8f4-4f68-8ff0-2c7856ff4a7c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.894464] env[61987]: DEBUG oslo_vmware.api [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 986.894464] env[61987]: value = "task-1062094" [ 986.894464] env[61987]: _type = "Task" [ 986.894464] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.906411] env[61987]: DEBUG oslo_vmware.api [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062094, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.955096] env[61987]: DEBUG nova.objects.instance [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lazy-loading 'flavor' on Instance uuid ae1396b3-c28b-4d1c-9a3a-3dc2170847ca {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 986.967180] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]528ff4b6-67f9-1779-57a3-07f1cde374bb, 'name': SearchDatastore_Task, 'duration_secs': 0.027324} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.968056] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6132fcf2-99d4-463a-b091-e19405291b92 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.976161] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 986.976161] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52535088-5899-2ec8-bc75-a2e5a432d458" [ 986.976161] env[61987]: _type = "Task" [ 986.976161] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.976161] env[61987]: DEBUG nova.network.neutron [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Successfully updated port: e9aa9f0a-f5e5-4d71-bddd-c817a7b4baf1 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 986.990497] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52535088-5899-2ec8-bc75-a2e5a432d458, 'name': SearchDatastore_Task, 'duration_secs': 0.010516} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.991525] env[61987]: DEBUG oslo_concurrency.lockutils [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.991801] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 473e0b73-6fcb-44c1-abf7-47d30c58af1b/473e0b73-6fcb-44c1-abf7-47d30c58af1b.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 986.992154] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-662deafb-7d8e-4333-9e31-7b7bb169b9f2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.998391] env[61987]: INFO nova.network.neutron [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Port b42049a4-5eda-4603-950e-36103e3b012a from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 986.998778] env[61987]: DEBUG nova.network.neutron [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Updating instance_info_cache with network_info: [{"id": "c4833bbc-333e-46c2-a140-5e8d493d3973", "address": "fa:16:3e:11:e5:35", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4833bbc-33", "ovs_interfaceid": "c4833bbc-333e-46c2-a140-5e8d493d3973", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.002585] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 987.002585] env[61987]: value = "task-1062095" [ 987.002585] env[61987]: _type = "Task" [ 987.002585] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.012281] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062095, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.053415] env[61987]: DEBUG nova.compute.manager [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 987.053948] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 987.054914] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07a616e9-cbc6-4ff0-9cb6-9aef412a2141 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.063529] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 987.064392] env[61987]: DEBUG nova.network.neutron [None req-4750d3ba-3203-4f69-a7e3-317423f1424f tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 987.068712] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01c1a88d-e5b1-46c7-9711-be23503193b4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.072744] env[61987]: DEBUG oslo_vmware.api [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 987.072744] env[61987]: value = "task-1062096" [ 987.072744] env[61987]: _type = "Task" [ 987.072744] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.084743] env[61987]: DEBUG oslo_vmware.api [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062096, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.196019] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062093, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.310143] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.310143] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.310143] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.310143] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61987) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 987.311030] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db838b49-293b-42fc-a1ef-771f7f14f9f9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.320297] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7212fd9-6a70-44bd-924f-6a40ce56a79b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.337428] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3320cd1e-a231-4bdc-8160-c314fcfdac2d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.345255] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c76927c3-f8af-4947-95a3-732809df2428 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.377964] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180189MB free_disk=179GB free_vcpus=48 pci_devices=None {{(pid=61987) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 987.378206] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.378352] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.406278] env[61987]: DEBUG oslo_vmware.api [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062094, 'name': ReconfigVM_Task, 'duration_secs': 0.199488} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.406597] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Updating instance '4dc624ca-c23c-4c4f-8481-45f954e1c068' progress to 33 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 987.422100] env[61987]: DEBUG nova.network.neutron [-] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.463648] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f337ed75-4bb1-4a10-9ab2-79c51eacee53 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.795s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.479104] env[61987]: DEBUG oslo_concurrency.lockutils [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquiring lock "refresh_cache-e6236bf1-4068-4a6e-9787-cd812381e122" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.479306] env[61987]: DEBUG oslo_concurrency.lockutils [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquired lock "refresh_cache-e6236bf1-4068-4a6e-9787-cd812381e122" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.479427] env[61987]: DEBUG nova.network.neutron [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 987.501751] env[61987]: DEBUG oslo_concurrency.lockutils [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "refresh_cache-efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.514019] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062095, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.514066} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.514820] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 473e0b73-6fcb-44c1-abf7-47d30c58af1b/473e0b73-6fcb-44c1-abf7-47d30c58af1b.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 987.515070] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 987.515386] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ceed8ad9-2daa-4bf4-8ce0-ab29432760ba {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.521575] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 987.521575] env[61987]: value = "task-1062097" [ 987.521575] env[61987]: _type = "Task" [ 987.521575] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.529370] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062097, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.581254] env[61987]: DEBUG oslo_vmware.api [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062096, 'name': PowerOffVM_Task, 'duration_secs': 0.26726} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.581604] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 987.581784] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 987.582653] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ad733c09-aefc-40d6-8304-54208b943910 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.701346] env[61987]: DEBUG oslo_vmware.api [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062093, 'name': PowerOnVM_Task, 'duration_secs': 0.558207} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.704076] env[61987]: DEBUG oslo_concurrency.lockutils [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "865222ae-12da-4e5a-84af-041003b50520" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.704076] env[61987]: DEBUG oslo_concurrency.lockutils [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "865222ae-12da-4e5a-84af-041003b50520" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.705288] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 987.705503] env[61987]: INFO nova.compute.manager [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Took 8.27 seconds to spawn the instance on the hypervisor. [ 987.705689] env[61987]: DEBUG nova.compute.manager [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 987.706319] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 987.706521] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 987.706709] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Deleting the datastore file [datastore1] efdf62cc-9a3f-4b6b-a119-29a8096848ee {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 987.707988] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d21e9914-faf4-408a-bdb0-45d06efc58e5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.710873] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-04208537-6015-4779-8289-0c6be337d83c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.720309] env[61987]: DEBUG oslo_vmware.api [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 987.720309] env[61987]: value = "task-1062099" [ 987.720309] env[61987]: _type = "Task" [ 987.720309] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.730358] env[61987]: DEBUG oslo_vmware.api [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062099, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.835229] env[61987]: DEBUG nova.network.neutron [None req-4750d3ba-3203-4f69-a7e3-317423f1424f tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Updating instance_info_cache with network_info: [{"id": "feaf0c72-3782-4054-907b-f98cc72cd02a", "address": "fa:16:3e:13:f5:64", "network": {"id": "b45b790f-b724-4a9a-b54b-8f8319b13a88", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1186325730-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7f68eb92e124eab9beb7b60ddcc1df3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfeaf0c72-37", "ovs_interfaceid": "feaf0c72-3782-4054-907b-f98cc72cd02a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.854023] env[61987]: DEBUG nova.compute.manager [req-3514d1d7-1db3-44ca-b766-c7e826d4e291 req-82cf8775-ce6b-449c-b392-aff86a13d865 service nova] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Received event network-vif-deleted-e237b976-f4a7-4dd6-8003-9b80e7f8e5fe {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 987.854280] env[61987]: DEBUG nova.compute.manager [req-3514d1d7-1db3-44ca-b766-c7e826d4e291 req-82cf8775-ce6b-449c-b392-aff86a13d865 service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Received event network-changed-feaf0c72-3782-4054-907b-f98cc72cd02a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 987.854450] env[61987]: DEBUG nova.compute.manager [req-3514d1d7-1db3-44ca-b766-c7e826d4e291 req-82cf8775-ce6b-449c-b392-aff86a13d865 service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Refreshing instance network info cache due to event network-changed-feaf0c72-3782-4054-907b-f98cc72cd02a. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 987.854646] env[61987]: DEBUG oslo_concurrency.lockutils [req-3514d1d7-1db3-44ca-b766-c7e826d4e291 req-82cf8775-ce6b-449c-b392-aff86a13d865 service nova] Acquiring lock "refresh_cache-cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.908267] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.908541] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.912900] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 987.913150] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 987.913319] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 987.913515] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 987.913675] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 987.913820] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 987.914090] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 987.914210] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 987.914381] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 987.914553] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 987.914735] env[61987]: DEBUG nova.virt.hardware [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 987.920282] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Reconfiguring VM instance instance-00000061 to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 987.921344] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e10aa61d-829a-45a0-b268-741ba3adcded {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.935971] env[61987]: INFO nova.compute.manager [-] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Took 1.55 seconds to deallocate network for instance. [ 987.943083] env[61987]: DEBUG oslo_vmware.api [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 987.943083] env[61987]: value = "task-1062100" [ 987.943083] env[61987]: _type = "Task" [ 987.943083] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.951659] env[61987]: DEBUG oslo_vmware.api [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062100, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.009161] env[61987]: DEBUG oslo_concurrency.lockutils [None req-18cf497f-8bc1-4383-a7bf-d88c57fe8509 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "interface-efdf62cc-9a3f-4b6b-a119-29a8096848ee-697d9636-5a3c-4b4e-8a86-3ea919b04757" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.013942] env[61987]: DEBUG nova.network.neutron [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 988.033079] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062097, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059875} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.035358] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 988.036201] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d256a66-9b56-4d74-825b-6cc23ada3eec {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.058956] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 473e0b73-6fcb-44c1-abf7-47d30c58af1b/473e0b73-6fcb-44c1-abf7-47d30c58af1b.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 988.061326] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5470493-c140-4336-b2d6-01a612990c5d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.082280] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 988.082280] env[61987]: value = "task-1062101" [ 988.082280] env[61987]: _type = "Task" [ 988.082280] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.090142] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062101, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.160520] env[61987]: DEBUG nova.network.neutron [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Updating instance_info_cache with network_info: [{"id": "e9aa9f0a-f5e5-4d71-bddd-c817a7b4baf1", "address": "fa:16:3e:87:a2:7d", "network": {"id": "1157ea83-0804-47d3-96b7-7a5a16809f1a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-318208769-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "630439422a354de0b32ac5eefaa990b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69744f59-ecac-4b0b-831e-82a274d7acbb", "external-id": "nsx-vlan-transportzone-770", "segmentation_id": 770, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9aa9f0a-f5", "ovs_interfaceid": "e9aa9f0a-f5e5-4d71-bddd-c817a7b4baf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.208604] env[61987]: DEBUG nova.compute.utils [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 988.230218] env[61987]: INFO nova.compute.manager [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Took 13.12 seconds to build instance. [ 988.236768] env[61987]: DEBUG oslo_vmware.api [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062099, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.338105] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4750d3ba-3203-4f69-a7e3-317423f1424f tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Releasing lock "refresh_cache-cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.338449] env[61987]: DEBUG nova.compute.manager [None req-4750d3ba-3203-4f69-a7e3-317423f1424f tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Inject network info {{(pid=61987) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7692}} [ 988.338764] env[61987]: DEBUG nova.compute.manager [None req-4750d3ba-3203-4f69-a7e3-317423f1424f tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] network_info to inject: |[{"id": "feaf0c72-3782-4054-907b-f98cc72cd02a", "address": "fa:16:3e:13:f5:64", "network": {"id": "b45b790f-b724-4a9a-b54b-8f8319b13a88", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1186325730-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7f68eb92e124eab9beb7b60ddcc1df3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfeaf0c72-37", "ovs_interfaceid": "feaf0c72-3782-4054-907b-f98cc72cd02a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7693}} [ 988.346393] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4750d3ba-3203-4f69-a7e3-317423f1424f tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Reconfiguring VM instance to set the machine id {{(pid=61987) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 988.346706] env[61987]: DEBUG oslo_concurrency.lockutils [req-3514d1d7-1db3-44ca-b766-c7e826d4e291 req-82cf8775-ce6b-449c-b392-aff86a13d865 service nova] Acquired lock "refresh_cache-cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.346892] env[61987]: DEBUG nova.network.neutron [req-3514d1d7-1db3-44ca-b766-c7e826d4e291 req-82cf8775-ce6b-449c-b392-aff86a13d865 service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Refreshing network info cache for port feaf0c72-3782-4054-907b-f98cc72cd02a {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 988.348070] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45971a05-c691-455c-a2a9-c32ebaba8de8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.365417] env[61987]: DEBUG oslo_vmware.api [None req-4750d3ba-3203-4f69-a7e3-317423f1424f tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Waiting for the task: (returnval){ [ 988.365417] env[61987]: value = "task-1062102" [ 988.365417] env[61987]: _type = "Task" [ 988.365417] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.373763] env[61987]: DEBUG oslo_vmware.api [None req-4750d3ba-3203-4f69-a7e3-317423f1424f tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062102, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.389077] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Applying migration context for instance 4dc624ca-c23c-4c4f-8481-45f954e1c068 as it has an incoming, in-progress migration a799cdd4-fc8b-40af-a2f0-b591c5463c7d. Migration status is migrating {{(pid=61987) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 988.390644] env[61987]: INFO nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Updating resource usage from migration a799cdd4-fc8b-40af-a2f0-b591c5463c7d [ 988.411405] env[61987]: DEBUG nova.compute.utils [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 988.415366] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 84a4a69b-04db-4546-ba89-9039e382a0c4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 988.415509] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 69a22e40-d469-4500-926e-0a12a233f252 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 988.415634] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance efdf62cc-9a3f-4b6b-a119-29a8096848ee actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 988.415753] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance ae1396b3-c28b-4d1c-9a3a-3dc2170847ca actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 988.415873] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 988.415989] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 865222ae-12da-4e5a-84af-041003b50520 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 988.416129] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 4d15848d-5265-4897-bfa7-f3ad4c02ed83 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 988.416247] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 6ec74ee9-a73b-4eac-97f2-abac3b1e3191 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 988.416360] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 988.416473] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 473e0b73-6fcb-44c1-abf7-47d30c58af1b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 988.416586] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Migration a799cdd4-fc8b-40af-a2f0-b591c5463c7d is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 988.416698] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 4dc624ca-c23c-4c4f-8481-45f954e1c068 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 988.416810] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance e6236bf1-4068-4a6e-9787-cd812381e122 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 988.417015] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 988.417172] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3072MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 988.442340] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.455826] env[61987]: DEBUG oslo_vmware.api [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062100, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.591964] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062101, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.593754] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf2bc6d-ce63-4496-95b5-3cd30af64ed2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.601505] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3bb8533-cf4c-43fa-a676-36a3871f61d8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.632498] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a28ceb6-ec9e-484d-bee4-b8fb64f1a480 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.639855] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c569b63c-7946-4da2-ad42-7354da87d22a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.652507] env[61987]: DEBUG nova.compute.provider_tree [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 988.663370] env[61987]: DEBUG oslo_concurrency.lockutils [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Releasing lock "refresh_cache-e6236bf1-4068-4a6e-9787-cd812381e122" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.663666] env[61987]: DEBUG nova.compute.manager [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Instance network_info: |[{"id": "e9aa9f0a-f5e5-4d71-bddd-c817a7b4baf1", "address": "fa:16:3e:87:a2:7d", "network": {"id": "1157ea83-0804-47d3-96b7-7a5a16809f1a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-318208769-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "630439422a354de0b32ac5eefaa990b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69744f59-ecac-4b0b-831e-82a274d7acbb", "external-id": "nsx-vlan-transportzone-770", "segmentation_id": 770, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9aa9f0a-f5", "ovs_interfaceid": "e9aa9f0a-f5e5-4d71-bddd-c817a7b4baf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 988.664074] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:a2:7d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69744f59-ecac-4b0b-831e-82a274d7acbb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e9aa9f0a-f5e5-4d71-bddd-c817a7b4baf1', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 988.671457] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Creating folder: Project (630439422a354de0b32ac5eefaa990b9). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 988.671954] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2f40c4ad-fc8e-4a64-b935-5a411cad2e7e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.682331] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Created folder: Project (630439422a354de0b32ac5eefaa990b9) in parent group-v234219. [ 988.682528] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Creating folder: Instances. Parent ref: group-v234408. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 988.682767] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aa2dde34-feae-4122-b5f7-61f4f197e78f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.692192] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Created folder: Instances in parent group-v234408. [ 988.692432] env[61987]: DEBUG oslo.service.loopingcall [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 988.692619] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 988.692823] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ff070784-f4c5-4292-aff4-cc1e617d06c8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.711276] env[61987]: DEBUG oslo_concurrency.lockutils [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "865222ae-12da-4e5a-84af-041003b50520" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.712822] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 988.712822] env[61987]: value = "task-1062105" [ 988.712822] env[61987]: _type = "Task" [ 988.712822] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.723473] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062105, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.733200] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64012953-d76e-4032-ba25-a1769e4cca60 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.629s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.733478] env[61987]: DEBUG oslo_vmware.api [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062099, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.765368} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.733761] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 988.733907] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 988.734164] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 988.734378] env[61987]: INFO nova.compute.manager [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Took 1.68 seconds to destroy the instance on the hypervisor. [ 988.734623] env[61987]: DEBUG oslo.service.loopingcall [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 988.734864] env[61987]: DEBUG nova.compute.manager [-] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 988.734973] env[61987]: DEBUG nova.network.neutron [-] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 988.776708] env[61987]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port b42049a4-5eda-4603-950e-36103e3b012a could not be found.", "detail": ""}} {{(pid=61987) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 988.776963] env[61987]: DEBUG nova.network.neutron [-] Unable to show port b42049a4-5eda-4603-950e-36103e3b012a as it no longer exists. {{(pid=61987) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 988.876274] env[61987]: DEBUG oslo_vmware.api [None req-4750d3ba-3203-4f69-a7e3-317423f1424f tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062102, 'name': ReconfigVM_Task, 'duration_secs': 0.195051} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.876636] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4750d3ba-3203-4f69-a7e3-317423f1424f tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Reconfigured VM instance to set the machine id {{(pid=61987) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 988.914030] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.953830] env[61987]: DEBUG oslo_vmware.api [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062100, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.093087] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062101, 'name': ReconfigVM_Task, 'duration_secs': 0.852402} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.093399] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 473e0b73-6fcb-44c1-abf7-47d30c58af1b/473e0b73-6fcb-44c1-abf7-47d30c58af1b.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 989.094057] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c8aabd63-2063-4265-b1d8-0dcd4d3d2689 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.099947] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 989.099947] env[61987]: value = "task-1062106" [ 989.099947] env[61987]: _type = "Task" [ 989.099947] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.108115] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062106, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.155375] env[61987]: DEBUG nova.scheduler.client.report [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 989.225204] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062105, 'name': CreateVM_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.300709] env[61987]: DEBUG nova.compute.manager [req-1aaedf00-d205-46ca-be5f-afc4a83781d5 req-b998ffda-2948-4ffd-ae76-9fc091f5a764 service nova] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Received event network-vif-plugged-e9aa9f0a-f5e5-4d71-bddd-c817a7b4baf1 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 989.300989] env[61987]: DEBUG oslo_concurrency.lockutils [req-1aaedf00-d205-46ca-be5f-afc4a83781d5 req-b998ffda-2948-4ffd-ae76-9fc091f5a764 service nova] Acquiring lock "e6236bf1-4068-4a6e-9787-cd812381e122-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.301245] env[61987]: DEBUG oslo_concurrency.lockutils [req-1aaedf00-d205-46ca-be5f-afc4a83781d5 req-b998ffda-2948-4ffd-ae76-9fc091f5a764 service nova] Lock "e6236bf1-4068-4a6e-9787-cd812381e122-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.301477] env[61987]: DEBUG oslo_concurrency.lockutils [req-1aaedf00-d205-46ca-be5f-afc4a83781d5 req-b998ffda-2948-4ffd-ae76-9fc091f5a764 service nova] Lock "e6236bf1-4068-4a6e-9787-cd812381e122-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.301653] env[61987]: DEBUG nova.compute.manager [req-1aaedf00-d205-46ca-be5f-afc4a83781d5 req-b998ffda-2948-4ffd-ae76-9fc091f5a764 service nova] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] No waiting events found dispatching network-vif-plugged-e9aa9f0a-f5e5-4d71-bddd-c817a7b4baf1 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 989.301840] env[61987]: WARNING nova.compute.manager [req-1aaedf00-d205-46ca-be5f-afc4a83781d5 req-b998ffda-2948-4ffd-ae76-9fc091f5a764 service nova] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Received unexpected event network-vif-plugged-e9aa9f0a-f5e5-4d71-bddd-c817a7b4baf1 for instance with vm_state building and task_state spawning. [ 989.302021] env[61987]: DEBUG nova.compute.manager [req-1aaedf00-d205-46ca-be5f-afc4a83781d5 req-b998ffda-2948-4ffd-ae76-9fc091f5a764 service nova] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Received event network-changed-e9aa9f0a-f5e5-4d71-bddd-c817a7b4baf1 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 989.302639] env[61987]: DEBUG nova.compute.manager [req-1aaedf00-d205-46ca-be5f-afc4a83781d5 req-b998ffda-2948-4ffd-ae76-9fc091f5a764 service nova] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Refreshing instance network info cache due to event network-changed-e9aa9f0a-f5e5-4d71-bddd-c817a7b4baf1. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 989.303213] env[61987]: DEBUG oslo_concurrency.lockutils [req-1aaedf00-d205-46ca-be5f-afc4a83781d5 req-b998ffda-2948-4ffd-ae76-9fc091f5a764 service nova] Acquiring lock "refresh_cache-e6236bf1-4068-4a6e-9787-cd812381e122" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.303213] env[61987]: DEBUG oslo_concurrency.lockutils [req-1aaedf00-d205-46ca-be5f-afc4a83781d5 req-b998ffda-2948-4ffd-ae76-9fc091f5a764 service nova] Acquired lock "refresh_cache-e6236bf1-4068-4a6e-9787-cd812381e122" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.303459] env[61987]: DEBUG nova.network.neutron [req-1aaedf00-d205-46ca-be5f-afc4a83781d5 req-b998ffda-2948-4ffd-ae76-9fc091f5a764 service nova] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Refreshing network info cache for port e9aa9f0a-f5e5-4d71-bddd-c817a7b4baf1 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 989.433667] env[61987]: DEBUG nova.network.neutron [req-3514d1d7-1db3-44ca-b766-c7e826d4e291 req-82cf8775-ce6b-449c-b392-aff86a13d865 service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Updated VIF entry in instance network info cache for port feaf0c72-3782-4054-907b-f98cc72cd02a. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 989.434149] env[61987]: DEBUG nova.network.neutron [req-3514d1d7-1db3-44ca-b766-c7e826d4e291 req-82cf8775-ce6b-449c-b392-aff86a13d865 service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Updating instance_info_cache with network_info: [{"id": "feaf0c72-3782-4054-907b-f98cc72cd02a", "address": "fa:16:3e:13:f5:64", "network": {"id": "b45b790f-b724-4a9a-b54b-8f8319b13a88", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1186325730-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e7f68eb92e124eab9beb7b60ddcc1df3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "767a3a48-41d4-4a0c-961d-0024837f63bd", "external-id": "nsx-vlan-transportzone-141", "segmentation_id": 141, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfeaf0c72-37", "ovs_interfaceid": "feaf0c72-3782-4054-907b-f98cc72cd02a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.454528] env[61987]: DEBUG oslo_vmware.api [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062100, 'name': ReconfigVM_Task, 'duration_secs': 1.333274} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.454822] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Reconfigured VM instance instance-00000061 to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 989.455686] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56da430a-f190-4b62-a668-af4eab0b34dd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.479298] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 4dc624ca-c23c-4c4f-8481-45f954e1c068/4dc624ca-c23c-4c4f-8481-45f954e1c068.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 989.479298] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f716cd73-17c9-4ac0-b8d7-b6d307b6bcbb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.504163] env[61987]: DEBUG oslo_vmware.api [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 989.504163] env[61987]: value = "task-1062107" [ 989.504163] env[61987]: _type = "Task" [ 989.504163] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.516565] env[61987]: DEBUG oslo_vmware.api [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062107, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.610655] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062106, 'name': Rename_Task, 'duration_secs': 0.185811} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.610888] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 989.611123] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a70bac4e-2fc5-41c0-a2b3-40af5527a4ef {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.617652] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 989.617652] env[61987]: value = "task-1062108" [ 989.617652] env[61987]: _type = "Task" [ 989.617652] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.625367] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062108, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.659866] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61987) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 989.660166] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.282s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.660544] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.218s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.660766] env[61987]: DEBUG nova.objects.instance [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lazy-loading 'resources' on Instance uuid 6ec74ee9-a73b-4eac-97f2-abac3b1e3191 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 989.728250] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062105, 'name': CreateVM_Task, 'duration_secs': 0.793161} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.728465] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 989.729755] env[61987]: DEBUG oslo_concurrency.lockutils [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.729755] env[61987]: DEBUG oslo_concurrency.lockutils [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.730179] env[61987]: DEBUG oslo_concurrency.lockutils [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 989.730490] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6443eae-cebc-4b5c-b573-d3c6991069c8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.735979] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 989.735979] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52f18090-4efb-d136-5781-3a8c7b1f9095" [ 989.735979] env[61987]: _type = "Task" [ 989.735979] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.745638] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52f18090-4efb-d136-5781-3a8c7b1f9095, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.776949] env[61987]: DEBUG oslo_concurrency.lockutils [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "865222ae-12da-4e5a-84af-041003b50520" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.777267] env[61987]: DEBUG oslo_concurrency.lockutils [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "865222ae-12da-4e5a-84af-041003b50520" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.777520] env[61987]: INFO nova.compute.manager [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Attaching volume 87789bc2-c4d6-4c37-ba7b-6211d2bb6126 to /dev/sdb [ 989.819474] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-428b5eb7-cf04-4d58-90c6-debaffa90357 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.826922] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f81159f4-2e16-43ff-871f-13c5551d6ee9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.840124] env[61987]: DEBUG nova.virt.block_device [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Updating existing volume attachment record: e8064854-49a8-488b-a135-944629a093c3 {{(pid=61987) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 989.886298] env[61987]: DEBUG nova.compute.manager [req-7f37e270-8d2d-40b1-8477-e6074292cd3a req-44d1f2c2-cdcd-4fdf-9fec-c90c2c1b98c2 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Received event network-vif-deleted-c4833bbc-333e-46c2-a140-5e8d493d3973 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 989.886565] env[61987]: INFO nova.compute.manager [req-7f37e270-8d2d-40b1-8477-e6074292cd3a req-44d1f2c2-cdcd-4fdf-9fec-c90c2c1b98c2 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Neutron deleted interface c4833bbc-333e-46c2-a140-5e8d493d3973; detaching it from the instance and deleting it from the info cache [ 989.886753] env[61987]: DEBUG nova.network.neutron [req-7f37e270-8d2d-40b1-8477-e6074292cd3a req-44d1f2c2-cdcd-4fdf-9fec-c90c2c1b98c2 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.937761] env[61987]: DEBUG oslo_concurrency.lockutils [req-3514d1d7-1db3-44ca-b766-c7e826d4e291 req-82cf8775-ce6b-449c-b392-aff86a13d865 service nova] Releasing lock "refresh_cache-cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.001744] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.001744] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.001744] env[61987]: INFO nova.compute.manager [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Attaching volume 9d3822d3-6b17-4fad-84da-5f273eb187c1 to /dev/sdc [ 990.016966] env[61987]: DEBUG oslo_vmware.api [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062107, 'name': ReconfigVM_Task, 'duration_secs': 0.273808} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.017300] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 4dc624ca-c23c-4c4f-8481-45f954e1c068/4dc624ca-c23c-4c4f-8481-45f954e1c068.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 990.017704] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Updating instance '4dc624ca-c23c-4c4f-8481-45f954e1c068' progress to 50 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 990.022164] env[61987]: DEBUG nova.network.neutron [-] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.038971] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90978062-38e7-494b-9be8-76fd2f2c3039 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.046773] env[61987]: DEBUG nova.network.neutron [req-1aaedf00-d205-46ca-be5f-afc4a83781d5 req-b998ffda-2948-4ffd-ae76-9fc091f5a764 service nova] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Updated VIF entry in instance network info cache for port e9aa9f0a-f5e5-4d71-bddd-c817a7b4baf1. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 990.047158] env[61987]: DEBUG nova.network.neutron [req-1aaedf00-d205-46ca-be5f-afc4a83781d5 req-b998ffda-2948-4ffd-ae76-9fc091f5a764 service nova] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Updating instance_info_cache with network_info: [{"id": "e9aa9f0a-f5e5-4d71-bddd-c817a7b4baf1", "address": "fa:16:3e:87:a2:7d", "network": {"id": "1157ea83-0804-47d3-96b7-7a5a16809f1a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-318208769-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "630439422a354de0b32ac5eefaa990b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69744f59-ecac-4b0b-831e-82a274d7acbb", "external-id": "nsx-vlan-transportzone-770", "segmentation_id": 770, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9aa9f0a-f5", "ovs_interfaceid": "e9aa9f0a-f5e5-4d71-bddd-c817a7b4baf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.048997] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56c3d1a8-e6ec-4ce0-9873-447e580f10b4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.067014] env[61987]: DEBUG nova.virt.block_device [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Updating existing volume attachment record: 504b6e65-c6bf-4756-97dc-6acba8fd53c8 {{(pid=61987) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 990.127919] env[61987]: DEBUG oslo_vmware.api [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062108, 'name': PowerOnVM_Task, 'duration_secs': 0.477177} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.128259] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 990.128563] env[61987]: INFO nova.compute.manager [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Took 8.33 seconds to spawn the instance on the hypervisor. [ 990.128747] env[61987]: DEBUG nova.compute.manager [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 990.129601] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93e69033-5f56-4ceb-bb94-5812ad6d9ea6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.223508] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Acquiring lock "cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.223847] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Lock "cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.224100] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Acquiring lock "cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.224296] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Lock "cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.224473] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Lock "cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.226673] env[61987]: INFO nova.compute.manager [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Terminating instance [ 990.248895] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52f18090-4efb-d136-5781-3a8c7b1f9095, 'name': SearchDatastore_Task, 'duration_secs': 0.009562} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.249596] env[61987]: DEBUG oslo_concurrency.lockutils [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.250084] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 990.251539] env[61987]: DEBUG oslo_concurrency.lockutils [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.251781] env[61987]: DEBUG oslo_concurrency.lockutils [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.251949] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 990.252239] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-df989748-f857-45da-bf57-d4b45a0843a3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.261163] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 990.261996] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 990.262098] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88d91f42-6202-4b99-8a64-998e24d04df9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.267207] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 990.267207] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5200a38f-ec07-ad19-82b8-a4a3689ee41b" [ 990.267207] env[61987]: _type = "Task" [ 990.267207] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.277707] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5200a38f-ec07-ad19-82b8-a4a3689ee41b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.339951] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65427625-c838-4804-a7d3-fc030fa1664a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.347624] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d84ba0-386e-44c9-b366-2c92c8789a12 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.381360] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3f95e2-5caf-4414-9e4a-91171e2b03c0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.389193] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a007a565-adfc-4aa8-9b57-0f5ce05eb83a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.395222] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4bd89875-ece2-4e78-914f-8bc151b70c80 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.406987] env[61987]: DEBUG nova.compute.provider_tree [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.412706] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e60aae-fcc0-467e-ac14-37f3b4aa247c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.446508] env[61987]: DEBUG nova.compute.manager [req-7f37e270-8d2d-40b1-8477-e6074292cd3a req-44d1f2c2-cdcd-4fdf-9fec-c90c2c1b98c2 service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Detach interface failed, port_id=c4833bbc-333e-46c2-a140-5e8d493d3973, reason: Instance efdf62cc-9a3f-4b6b-a119-29a8096848ee could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 990.525547] env[61987]: INFO nova.compute.manager [-] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Took 1.79 seconds to deallocate network for instance. [ 990.526529] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7140e08a-463d-4f62-be17-b05bc20806ee {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.550663] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4b05e4f-a2b6-4e01-b740-465c6b56e16e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.576975] env[61987]: DEBUG oslo_concurrency.lockutils [req-1aaedf00-d205-46ca-be5f-afc4a83781d5 req-b998ffda-2948-4ffd-ae76-9fc091f5a764 service nova] Releasing lock "refresh_cache-e6236bf1-4068-4a6e-9787-cd812381e122" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.578693] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Updating instance '4dc624ca-c23c-4c4f-8481-45f954e1c068' progress to 67 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 990.648722] env[61987]: INFO nova.compute.manager [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Took 15.04 seconds to build instance. [ 990.730520] env[61987]: DEBUG nova.compute.manager [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 990.730841] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 990.731851] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd938b3-a337-4d31-b63c-02fa0254920d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.739925] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 990.740189] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1aba37f3-bba3-4c2a-b70a-942a24f74f7d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.745884] env[61987]: DEBUG oslo_vmware.api [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Waiting for the task: (returnval){ [ 990.745884] env[61987]: value = "task-1062111" [ 990.745884] env[61987]: _type = "Task" [ 990.745884] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.755871] env[61987]: DEBUG oslo_vmware.api [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062111, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.777518] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5200a38f-ec07-ad19-82b8-a4a3689ee41b, 'name': SearchDatastore_Task, 'duration_secs': 0.010051} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.777879] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfc08e2b-f6fc-4ac6-9a55-2d5addb6bb61 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.783410] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 990.783410] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5266e52b-199e-3d1b-4ded-af0c0c4b16f5" [ 990.783410] env[61987]: _type = "Task" [ 990.783410] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.791500] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5266e52b-199e-3d1b-4ded-af0c0c4b16f5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.909704] env[61987]: DEBUG nova.scheduler.client.report [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 991.051446] env[61987]: DEBUG oslo_concurrency.lockutils [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.119177] env[61987]: DEBUG nova.network.neutron [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Port 1735c272-060e-4a5d-836a-af4e2eed480a binding to destination host cpu-1 is already ACTIVE {{(pid=61987) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 991.150466] env[61987]: DEBUG oslo_concurrency.lockutils [None req-222d6e1a-157e-4e4d-a4c2-0674aea17b28 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "473e0b73-6fcb-44c1-abf7-47d30c58af1b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.559s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.256051] env[61987]: DEBUG oslo_vmware.api [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062111, 'name': PowerOffVM_Task, 'duration_secs': 0.186738} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.256474] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 991.256534] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 991.256757] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-16166bf7-b788-479e-a79f-d511c0a0762e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.292684] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5266e52b-199e-3d1b-4ded-af0c0c4b16f5, 'name': SearchDatastore_Task, 'duration_secs': 0.010885} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.292978] env[61987]: DEBUG oslo_concurrency.lockutils [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.293291] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] e6236bf1-4068-4a6e-9787-cd812381e122/e6236bf1-4068-4a6e-9787-cd812381e122.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 991.293553] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c79cbf9d-8c83-4da5-9c5f-b3ff8d2edc79 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.299407] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 991.299407] env[61987]: value = "task-1062113" [ 991.299407] env[61987]: _type = "Task" [ 991.299407] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.306880] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062113, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.387609] env[61987]: INFO nova.compute.manager [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Rescuing [ 991.387899] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "refresh_cache-473e0b73-6fcb-44c1-abf7-47d30c58af1b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.388078] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquired lock "refresh_cache-473e0b73-6fcb-44c1-abf7-47d30c58af1b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.388258] env[61987]: DEBUG nova.network.neutron [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 991.414712] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.754s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.417013] env[61987]: DEBUG oslo_concurrency.lockutils [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.366s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.417275] env[61987]: DEBUG nova.objects.instance [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lazy-loading 'resources' on Instance uuid efdf62cc-9a3f-4b6b-a119-29a8096848ee {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.435273] env[61987]: INFO nova.scheduler.client.report [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Deleted allocations for instance 6ec74ee9-a73b-4eac-97f2-abac3b1e3191 [ 991.811151] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062113, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.848508] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 991.848760] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 991.849010] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Deleting the datastore file [datastore1] cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 991.849344] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cf736938-8bca-4eeb-aefd-b29cdcd0fb58 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.855909] env[61987]: DEBUG oslo_vmware.api [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Waiting for the task: (returnval){ [ 991.855909] env[61987]: value = "task-1062114" [ 991.855909] env[61987]: _type = "Task" [ 991.855909] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.864740] env[61987]: DEBUG oslo_vmware.api [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062114, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.944778] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cd061328-f15f-48fe-9ca3-8aafa02f9b1f tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "6ec74ee9-a73b-4eac-97f2-abac3b1e3191" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.189s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.148882] env[61987]: DEBUG oslo_concurrency.lockutils [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "4dc624ca-c23c-4c4f-8481-45f954e1c068-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.148882] env[61987]: DEBUG oslo_concurrency.lockutils [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "4dc624ca-c23c-4c4f-8481-45f954e1c068-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.148882] env[61987]: DEBUG oslo_concurrency.lockutils [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "4dc624ca-c23c-4c4f-8481-45f954e1c068-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.155895] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21485b0b-6dc0-4b98-a3e7-88250ab8ae9f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.167998] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d730ae75-ddee-4215-be97-faf501652ea1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.201933] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-287792bc-1702-4f1c-b0ca-212a2dd3e68e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.211045] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e7a0a29-a915-4161-a5ca-3188613ba609 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.227258] env[61987]: DEBUG nova.compute.provider_tree [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.243647] env[61987]: DEBUG nova.network.neutron [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Updating instance_info_cache with network_info: [{"id": "5e6a927f-3d25-42bc-86b7-39767ddfbd98", "address": "fa:16:3e:c8:53:17", "network": {"id": "4ed4fbdc-05b0-44f8-a3d5-b7b288abd5e0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2067313775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a234a62da0e498fbe613fbcaaec3201", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e6a927f-3d", "ovs_interfaceid": "5e6a927f-3d25-42bc-86b7-39767ddfbd98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.310248] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062113, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.817492} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.310516] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] e6236bf1-4068-4a6e-9787-cd812381e122/e6236bf1-4068-4a6e-9787-cd812381e122.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 992.310664] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 992.310901] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6f9f9eb3-31ae-43e1-af3d-6a4f614c7ebe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.319892] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 992.319892] env[61987]: value = "task-1062117" [ 992.319892] env[61987]: _type = "Task" [ 992.319892] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.328116] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062117, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.366420] env[61987]: DEBUG oslo_vmware.api [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Task: {'id': task-1062114, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.300281} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.366514] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 992.366683] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 992.367295] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 992.367295] env[61987]: INFO nova.compute.manager [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Took 1.64 seconds to destroy the instance on the hypervisor. [ 992.367396] env[61987]: DEBUG oslo.service.loopingcall [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 992.367619] env[61987]: DEBUG nova.compute.manager [-] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 992.367705] env[61987]: DEBUG nova.network.neutron [-] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 992.730578] env[61987]: DEBUG nova.scheduler.client.report [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 992.745937] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Releasing lock "refresh_cache-473e0b73-6fcb-44c1-abf7-47d30c58af1b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.826305] env[61987]: DEBUG nova.compute.manager [req-0b8adc4b-3ecd-42fe-945a-a747c9245d97 req-772cec9d-fc66-426f-b2f3-9edf44315eab service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Received event network-vif-deleted-feaf0c72-3782-4054-907b-f98cc72cd02a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 992.826530] env[61987]: INFO nova.compute.manager [req-0b8adc4b-3ecd-42fe-945a-a747c9245d97 req-772cec9d-fc66-426f-b2f3-9edf44315eab service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Neutron deleted interface feaf0c72-3782-4054-907b-f98cc72cd02a; detaching it from the instance and deleting it from the info cache [ 992.826714] env[61987]: DEBUG nova.network.neutron [req-0b8adc4b-3ecd-42fe-945a-a747c9245d97 req-772cec9d-fc66-426f-b2f3-9edf44315eab service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.833225] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062117, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060027} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.833501] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 992.834349] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91544c5b-859b-4667-bc46-7c5f81836af0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.858236] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] e6236bf1-4068-4a6e-9787-cd812381e122/e6236bf1-4068-4a6e-9787-cd812381e122.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 992.858872] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5fa2cbda-a3c5-4113-8039-ca8c52953f7b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.877994] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "54d9bbcb-d01c-4e3b-8feb-4f8942d475d4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.878244] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "54d9bbcb-d01c-4e3b-8feb-4f8942d475d4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.882665] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 992.882665] env[61987]: value = "task-1062118" [ 992.882665] env[61987]: _type = "Task" [ 992.882665] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.891770] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062118, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.203651] env[61987]: DEBUG oslo_concurrency.lockutils [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "refresh_cache-4dc624ca-c23c-4c4f-8481-45f954e1c068" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.203865] env[61987]: DEBUG oslo_concurrency.lockutils [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired lock "refresh_cache-4dc624ca-c23c-4c4f-8481-45f954e1c068" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.204065] env[61987]: DEBUG nova.network.neutron [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 993.235546] env[61987]: DEBUG oslo_concurrency.lockutils [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.818s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.258743] env[61987]: INFO nova.scheduler.client.report [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Deleted allocations for instance efdf62cc-9a3f-4b6b-a119-29a8096848ee [ 993.298505] env[61987]: DEBUG nova.network.neutron [-] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.330605] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5c44037c-9b08-40e3-bea8-25176df4d941 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.342863] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-429c46a8-b98e-4e9f-b253-79f4539e54b8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.372426] env[61987]: DEBUG nova.compute.manager [req-0b8adc4b-3ecd-42fe-945a-a747c9245d97 req-772cec9d-fc66-426f-b2f3-9edf44315eab service nova] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Detach interface failed, port_id=feaf0c72-3782-4054-907b-f98cc72cd02a, reason: Instance cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 993.380482] env[61987]: DEBUG nova.compute.manager [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 993.391839] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062118, 'name': ReconfigVM_Task, 'duration_secs': 0.285987} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.392104] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Reconfigured VM instance instance-00000065 to attach disk [datastore2] e6236bf1-4068-4a6e-9787-cd812381e122/e6236bf1-4068-4a6e-9787-cd812381e122.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 993.392721] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-921cae8f-b25e-451f-9369-2e3b448a7c44 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.399118] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 993.399118] env[61987]: value = "task-1062119" [ 993.399118] env[61987]: _type = "Task" [ 993.399118] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.407046] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062119, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.766293] env[61987]: DEBUG oslo_concurrency.lockutils [None req-42873243-ba33-4003-b948-67ffaa5e81c2 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.221s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.767222] env[61987]: DEBUG oslo_concurrency.lockutils [req-48cad022-0bc4-4d4d-a47c-d0a16e474ef4 req-a0414416-8458-4cb2-a3f9-3956f2c9677f service nova] Acquired lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.768023] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a011f0b-7abc-49a3-9c38-60f39dd357be {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.776902] env[61987]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 993.777095] env[61987]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=61987) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 993.779616] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-30082851-b401-4e5c-a593-d6a085a1b144 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.791158] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21cf07e1-8099-4c0d-8ab8-c5886c3240de {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.804109] env[61987]: INFO nova.compute.manager [-] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Took 1.44 seconds to deallocate network for instance. [ 993.826037] env[61987]: ERROR root [req-48cad022-0bc4-4d4d-a47c-d0a16e474ef4 req-a0414416-8458-4cb2-a3f9-3956f2c9677f service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-234380' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-234380' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-234380' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-234380'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-234380' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-234380' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-234380'}\n"]: nova.exception.InstanceNotFound: Instance efdf62cc-9a3f-4b6b-a119-29a8096848ee could not be found. [ 993.826290] env[61987]: DEBUG oslo_concurrency.lockutils [req-48cad022-0bc4-4d4d-a47c-d0a16e474ef4 req-a0414416-8458-4cb2-a3f9-3956f2c9677f service nova] Releasing lock "efdf62cc-9a3f-4b6b-a119-29a8096848ee" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.826521] env[61987]: DEBUG nova.compute.manager [req-48cad022-0bc4-4d4d-a47c-d0a16e474ef4 req-a0414416-8458-4cb2-a3f9-3956f2c9677f service nova] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Detach interface failed, port_id=b42049a4-5eda-4603-950e-36103e3b012a, reason: Instance efdf62cc-9a3f-4b6b-a119-29a8096848ee could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 993.905198] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.905541] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.907044] env[61987]: INFO nova.compute.claims [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 993.915092] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062119, 'name': Rename_Task, 'duration_secs': 0.15185} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.915348] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 993.915597] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-287d764b-bed6-4634-9dff-5b61cffb1ea9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.923273] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 993.923273] env[61987]: value = "task-1062120" [ 993.923273] env[61987]: _type = "Task" [ 993.923273] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.927212] env[61987]: DEBUG nova.network.neutron [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Updating instance_info_cache with network_info: [{"id": "1735c272-060e-4a5d-836a-af4e2eed480a", "address": "fa:16:3e:69:d8:56", "network": {"id": "ca68bf96-3f18-45ee-80fb-64a1260ed996", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-733595377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd47eb44bc334bf3ae5813905903ecbc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a27fd90b-16a5-43af-bede-ae36762ece00", "external-id": "nsx-vlan-transportzone-197", "segmentation_id": 197, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1735c272-06", "ovs_interfaceid": "1735c272-060e-4a5d-836a-af4e2eed480a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.933058] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062120, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.304306] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 994.304637] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d7f80b78-31fc-4a96-abb0-519f5ff38e24 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.310762] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.312345] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 994.312345] env[61987]: value = "task-1062121" [ 994.312345] env[61987]: _type = "Task" [ 994.312345] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.321631] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062121, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.385938] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Volume attach. Driver type: vmdk {{(pid=61987) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 994.386209] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234411', 'volume_id': '87789bc2-c4d6-4c37-ba7b-6211d2bb6126', 'name': 'volume-87789bc2-c4d6-4c37-ba7b-6211d2bb6126', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '865222ae-12da-4e5a-84af-041003b50520', 'attached_at': '', 'detached_at': '', 'volume_id': '87789bc2-c4d6-4c37-ba7b-6211d2bb6126', 'serial': '87789bc2-c4d6-4c37-ba7b-6211d2bb6126'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 994.387218] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2418a8c3-5aaa-41ec-a9a3-961d82609000 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.404804] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f4b398-4f11-43fe-9572-bd522e078ed3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.432602] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] volume-87789bc2-c4d6-4c37-ba7b-6211d2bb6126/volume-87789bc2-c4d6-4c37-ba7b-6211d2bb6126.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 994.434920] env[61987]: DEBUG oslo_concurrency.lockutils [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Releasing lock "refresh_cache-4dc624ca-c23c-4c4f-8481-45f954e1c068" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.438217] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53170c4e-3848-4894-a39d-6b71a05a3e15 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.468207] env[61987]: DEBUG oslo_vmware.api [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062120, 'name': PowerOnVM_Task, 'duration_secs': 0.449703} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.470195] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 994.470463] env[61987]: INFO nova.compute.manager [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Took 8.13 seconds to spawn the instance on the hypervisor. [ 994.470697] env[61987]: DEBUG nova.compute.manager [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 994.471079] env[61987]: DEBUG oslo_vmware.api [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 994.471079] env[61987]: value = "task-1062122" [ 994.471079] env[61987]: _type = "Task" [ 994.471079] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.471799] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e66d8a2-3d25-49c7-b11a-973a0a0d7959 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.488895] env[61987]: DEBUG oslo_vmware.api [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062122, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.619310] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Volume attach. Driver type: vmdk {{(pid=61987) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 994.619669] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234412', 'volume_id': '9d3822d3-6b17-4fad-84da-5f273eb187c1', 'name': 'volume-9d3822d3-6b17-4fad-84da-5f273eb187c1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ae1396b3-c28b-4d1c-9a3a-3dc2170847ca', 'attached_at': '', 'detached_at': '', 'volume_id': '9d3822d3-6b17-4fad-84da-5f273eb187c1', 'serial': '9d3822d3-6b17-4fad-84da-5f273eb187c1'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 994.620756] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-916c49fd-a7f7-4735-8f9c-8bdf690e0e43 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.639876] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a142b45f-3700-4782-9f1d-df28f0ffa741 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.671128] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] volume-9d3822d3-6b17-4fad-84da-5f273eb187c1/volume-9d3822d3-6b17-4fad-84da-5f273eb187c1.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 994.671846] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8173478d-b173-4533-9899-35f8a47c74b6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.691019] env[61987]: DEBUG oslo_vmware.api [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 994.691019] env[61987]: value = "task-1062123" [ 994.691019] env[61987]: _type = "Task" [ 994.691019] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.699380] env[61987]: DEBUG oslo_vmware.api [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062123, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.822881] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062121, 'name': PowerOffVM_Task, 'duration_secs': 0.319056} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.823239] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 994.824125] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30e02a9f-242a-469c-81cc-f7dee5418179 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.843163] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53a202c9-2d97-4740-b414-cb42e9dab7b0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.879420] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 994.879869] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1dee78d8-a5d9-43b6-8952-3b0842686abe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.887676] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 994.887676] env[61987]: value = "task-1062124" [ 994.887676] env[61987]: _type = "Task" [ 994.887676] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.896166] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] VM already powered off {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 994.896383] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 994.896639] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.896805] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.896993] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 994.897252] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a326b38d-0406-45c9-be4d-cf9f27a9c54a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.904992] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 994.905553] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 994.906306] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-211e9868-2f0b-44fa-a4fd-915a3cdd1a28 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.911528] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 994.911528] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]528519b8-af83-6816-98e6-88447af0f280" [ 994.911528] env[61987]: _type = "Task" [ 994.911528] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.918762] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]528519b8-af83-6816-98e6-88447af0f280, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.980536] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac3b90c0-c290-4cd3-92ac-ab44bd8a4050 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.012095] env[61987]: DEBUG oslo_vmware.api [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062122, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.015157] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-654b4e8a-e9f2-42d2-b742-2355997da77e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.018262] env[61987]: INFO nova.compute.manager [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Took 14.31 seconds to build instance. [ 995.025444] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Updating instance '4dc624ca-c23c-4c4f-8481-45f954e1c068' progress to 83 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 995.151673] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db7fe596-6ddf-4e70-84b0-43fccc3c6d33 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.161703] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96de6dc9-b933-48c0-b139-3a6f66e70da9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.212614] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d21d831-b0be-4d24-9267-d4cb9b4dce2e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.220765] env[61987]: DEBUG oslo_vmware.api [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062123, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.223915] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffc08435-9a06-489d-ac58-48e9d1b98c07 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.237942] env[61987]: DEBUG nova.compute.provider_tree [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.425074] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]528519b8-af83-6816-98e6-88447af0f280, 'name': SearchDatastore_Task, 'duration_secs': 0.009138} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.425943] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b80ed4c-ebce-4440-b758-15cfe1302f19 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.432435] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 995.432435] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52ba7632-1a30-c8d1-221b-8e9e904b56f7" [ 995.432435] env[61987]: _type = "Task" [ 995.432435] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.441193] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52ba7632-1a30-c8d1-221b-8e9e904b56f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.488088] env[61987]: DEBUG oslo_vmware.api [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062122, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.520815] env[61987]: DEBUG oslo_concurrency.lockutils [None req-76509304-7f2c-4289-b535-acfdbb7761fe tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lock "e6236bf1-4068-4a6e-9787-cd812381e122" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.824s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.532891] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 995.533238] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2037bbc7-de25-48d1-a9a5-844de2907204 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.539607] env[61987]: DEBUG oslo_vmware.api [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 995.539607] env[61987]: value = "task-1062125" [ 995.539607] env[61987]: _type = "Task" [ 995.539607] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.547090] env[61987]: DEBUG oslo_vmware.api [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062125, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.718759] env[61987]: DEBUG oslo_vmware.api [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062123, 'name': ReconfigVM_Task, 'duration_secs': 0.568747} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.719039] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Reconfigured VM instance instance-0000005a to attach disk [datastore2] volume-9d3822d3-6b17-4fad-84da-5f273eb187c1/volume-9d3822d3-6b17-4fad-84da-5f273eb187c1.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 995.724239] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db5bcf80-b240-4cc0-a661-514a8cf66f82 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.739078] env[61987]: DEBUG oslo_vmware.api [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 995.739078] env[61987]: value = "task-1062126" [ 995.739078] env[61987]: _type = "Task" [ 995.739078] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.742594] env[61987]: DEBUG nova.scheduler.client.report [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 995.750571] env[61987]: DEBUG oslo_vmware.api [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062126, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.835290] env[61987]: INFO nova.compute.manager [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Rescuing [ 995.835563] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquiring lock "refresh_cache-e6236bf1-4068-4a6e-9787-cd812381e122" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.835727] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquired lock "refresh_cache-e6236bf1-4068-4a6e-9787-cd812381e122" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.835915] env[61987]: DEBUG nova.network.neutron [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 995.942663] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52ba7632-1a30-c8d1-221b-8e9e904b56f7, 'name': SearchDatastore_Task, 'duration_secs': 0.00928} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.943167] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.943521] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 473e0b73-6fcb-44c1-abf7-47d30c58af1b/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk. {{(pid=61987) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 995.943871] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3dcb7445-67b6-478a-99e1-73fd96596168 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.957044] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 995.957044] env[61987]: value = "task-1062127" [ 995.957044] env[61987]: _type = "Task" [ 995.957044] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.965238] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062127, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.987246] env[61987]: DEBUG oslo_vmware.api [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062122, 'name': ReconfigVM_Task, 'duration_secs': 1.378402} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.987545] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Reconfigured VM instance instance-0000005e to attach disk [datastore2] volume-87789bc2-c4d6-4c37-ba7b-6211d2bb6126/volume-87789bc2-c4d6-4c37-ba7b-6211d2bb6126.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 995.992389] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-073e083b-37c1-44de-8657-99829b78b235 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.008676] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "7be2179d-68f7-4103-a8c1-b6e5b88b8706" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.009382] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "7be2179d-68f7-4103-a8c1-b6e5b88b8706" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.012976] env[61987]: DEBUG oslo_vmware.api [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 996.012976] env[61987]: value = "task-1062128" [ 996.012976] env[61987]: _type = "Task" [ 996.012976] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.024538] env[61987]: DEBUG oslo_vmware.api [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062128, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.050373] env[61987]: DEBUG oslo_vmware.api [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062125, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.248328] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.342s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.248909] env[61987]: DEBUG nova.compute.manager [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 996.253221] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.942s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.253740] env[61987]: DEBUG nova.objects.instance [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Lazy-loading 'resources' on Instance uuid cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 996.264605] env[61987]: DEBUG oslo_vmware.api [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062126, 'name': ReconfigVM_Task, 'duration_secs': 0.135335} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.265961] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234412', 'volume_id': '9d3822d3-6b17-4fad-84da-5f273eb187c1', 'name': 'volume-9d3822d3-6b17-4fad-84da-5f273eb187c1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ae1396b3-c28b-4d1c-9a3a-3dc2170847ca', 'attached_at': '', 'detached_at': '', 'volume_id': '9d3822d3-6b17-4fad-84da-5f273eb187c1', 'serial': '9d3822d3-6b17-4fad-84da-5f273eb187c1'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 996.467028] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062127, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.511702] env[61987]: DEBUG nova.compute.manager [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 996.525540] env[61987]: DEBUG oslo_vmware.api [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062128, 'name': ReconfigVM_Task, 'duration_secs': 0.218553} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.525540] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234411', 'volume_id': '87789bc2-c4d6-4c37-ba7b-6211d2bb6126', 'name': 'volume-87789bc2-c4d6-4c37-ba7b-6211d2bb6126', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '865222ae-12da-4e5a-84af-041003b50520', 'attached_at': '', 'detached_at': '', 'volume_id': '87789bc2-c4d6-4c37-ba7b-6211d2bb6126', 'serial': '87789bc2-c4d6-4c37-ba7b-6211d2bb6126'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 996.552093] env[61987]: DEBUG oslo_vmware.api [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062125, 'name': PowerOnVM_Task, 'duration_secs': 0.576067} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.554468] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 996.554667] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-57b257f4-8cbb-4f16-a74b-3fa6f82633b5 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Updating instance '4dc624ca-c23c-4c4f-8481-45f954e1c068' progress to 100 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 996.609653] env[61987]: DEBUG nova.network.neutron [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Updating instance_info_cache with network_info: [{"id": "e9aa9f0a-f5e5-4d71-bddd-c817a7b4baf1", "address": "fa:16:3e:87:a2:7d", "network": {"id": "1157ea83-0804-47d3-96b7-7a5a16809f1a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-318208769-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "630439422a354de0b32ac5eefaa990b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69744f59-ecac-4b0b-831e-82a274d7acbb", "external-id": "nsx-vlan-transportzone-770", "segmentation_id": 770, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape9aa9f0a-f5", "ovs_interfaceid": "e9aa9f0a-f5e5-4d71-bddd-c817a7b4baf1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.754346] env[61987]: DEBUG nova.compute.utils [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 996.758187] env[61987]: DEBUG nova.compute.manager [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 996.758390] env[61987]: DEBUG nova.network.neutron [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 996.802023] env[61987]: DEBUG nova.policy [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e16f557dfdb04406b3fc772b891f05a4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '22e62741abfb4471bbc7e11ae7d407ca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 996.912388] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c3af72-c1df-4cf5-9fc8-e928bd50c722 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.920052] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cadd1f5-6ba3-4d9b-8399-6235d814bf6c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.950766] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62a0fbb0-b296-432d-a47e-ccbd38d06b0a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.957777] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f4c3dbf-7e9e-476a-8b38-320ca1a40c63 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.973974] env[61987]: DEBUG nova.compute.provider_tree [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.978185] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062127, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.54656} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.978653] env[61987]: INFO nova.virt.vmwareapi.ds_util [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 473e0b73-6fcb-44c1-abf7-47d30c58af1b/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk. [ 996.979517] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de8ecc51-6c41-42ef-86fe-1231ee2cec19 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.004400] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 473e0b73-6fcb-44c1-abf7-47d30c58af1b/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 997.004968] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9af47c8-a449-4d04-996f-b4b2896d908a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.025870] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 997.025870] env[61987]: value = "task-1062129" [ 997.025870] env[61987]: _type = "Task" [ 997.025870] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.038851] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062129, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.040017] env[61987]: DEBUG nova.network.neutron [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Successfully created port: b3159788-bb87-42ad-b6e5-201ff2afe647 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 997.043189] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.112331] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Releasing lock "refresh_cache-e6236bf1-4068-4a6e-9787-cd812381e122" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.263546] env[61987]: DEBUG nova.compute.manager [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 997.307622] env[61987]: DEBUG nova.objects.instance [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lazy-loading 'flavor' on Instance uuid ae1396b3-c28b-4d1c-9a3a-3dc2170847ca {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.479865] env[61987]: DEBUG nova.scheduler.client.report [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 997.539395] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062129, 'name': ReconfigVM_Task, 'duration_secs': 0.356704} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.539656] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 473e0b73-6fcb-44c1-abf7-47d30c58af1b/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 997.540526] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a593964e-ff61-4899-913b-4fdc3bf07f0b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.576698] env[61987]: DEBUG nova.objects.instance [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lazy-loading 'flavor' on Instance uuid 865222ae-12da-4e5a-84af-041003b50520 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.577324] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2cefbd36-f202-46d6-9216-fa9336d51098 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.593247] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 997.593247] env[61987]: value = "task-1062130" [ 997.593247] env[61987]: _type = "Task" [ 997.593247] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.601584] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062130, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.812443] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ec3c350c-f227-431d-a617-11befef29085 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.813s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.985985] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.732s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.988618] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.945s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.989930] env[61987]: INFO nova.compute.claims [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 998.007688] env[61987]: INFO nova.scheduler.client.report [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Deleted allocations for instance cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0 [ 998.090187] env[61987]: DEBUG oslo_concurrency.lockutils [None req-03f86b3d-004b-4495-a946-67fd0dc3bcad tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "865222ae-12da-4e5a-84af-041003b50520" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.313s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.105981] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062130, 'name': ReconfigVM_Task, 'duration_secs': 0.167092} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.106373] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 998.106893] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d0440a7a-742e-4303-8570-35640c19814b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.113709] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 998.113709] env[61987]: value = "task-1062131" [ 998.113709] env[61987]: _type = "Task" [ 998.113709] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.123709] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062131, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.274082] env[61987]: DEBUG nova.compute.manager [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 998.306882] env[61987]: DEBUG nova.virt.hardware [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 998.306882] env[61987]: DEBUG nova.virt.hardware [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 998.307436] env[61987]: DEBUG nova.virt.hardware [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 998.307809] env[61987]: DEBUG nova.virt.hardware [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 998.308131] env[61987]: DEBUG nova.virt.hardware [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 998.309237] env[61987]: DEBUG nova.virt.hardware [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 998.309237] env[61987]: DEBUG nova.virt.hardware [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 998.309237] env[61987]: DEBUG nova.virt.hardware [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 998.309237] env[61987]: DEBUG nova.virt.hardware [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 998.309430] env[61987]: DEBUG nova.virt.hardware [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 998.309516] env[61987]: DEBUG nova.virt.hardware [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 998.310507] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d49b364-1005-488c-9503-14e794004206 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.318815] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aedd6661-584e-413d-93d4-6ab31699953f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.344049] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ab627140-eb2c-416b-a1ec-22e753a078e1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "865222ae-12da-4e5a-84af-041003b50520" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.344356] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ab627140-eb2c-416b-a1ec-22e753a078e1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "865222ae-12da-4e5a-84af-041003b50520" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.462201] env[61987]: DEBUG nova.compute.manager [req-a1a5ff54-4ba9-4836-b3f2-2623ca06921c req-101e37ba-7363-48e3-b12a-404c469187d5 service nova] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Received event network-vif-plugged-b3159788-bb87-42ad-b6e5-201ff2afe647 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 998.462626] env[61987]: DEBUG oslo_concurrency.lockutils [req-a1a5ff54-4ba9-4836-b3f2-2623ca06921c req-101e37ba-7363-48e3-b12a-404c469187d5 service nova] Acquiring lock "54d9bbcb-d01c-4e3b-8feb-4f8942d475d4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.462726] env[61987]: DEBUG oslo_concurrency.lockutils [req-a1a5ff54-4ba9-4836-b3f2-2623ca06921c req-101e37ba-7363-48e3-b12a-404c469187d5 service nova] Lock "54d9bbcb-d01c-4e3b-8feb-4f8942d475d4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.462923] env[61987]: DEBUG oslo_concurrency.lockutils [req-a1a5ff54-4ba9-4836-b3f2-2623ca06921c req-101e37ba-7363-48e3-b12a-404c469187d5 service nova] Lock "54d9bbcb-d01c-4e3b-8feb-4f8942d475d4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.463371] env[61987]: DEBUG nova.compute.manager [req-a1a5ff54-4ba9-4836-b3f2-2623ca06921c req-101e37ba-7363-48e3-b12a-404c469187d5 service nova] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] No waiting events found dispatching network-vif-plugged-b3159788-bb87-42ad-b6e5-201ff2afe647 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 998.463600] env[61987]: WARNING nova.compute.manager [req-a1a5ff54-4ba9-4836-b3f2-2623ca06921c req-101e37ba-7363-48e3-b12a-404c469187d5 service nova] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Received unexpected event network-vif-plugged-b3159788-bb87-42ad-b6e5-201ff2afe647 for instance with vm_state building and task_state spawning. [ 998.488469] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0334cc2a-9580-4a86-b414-c6d58fccaa2d tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.490102] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0334cc2a-9580-4a86-b414-c6d58fccaa2d tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.517086] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f3d24838-d565-45ad-942a-1f393221dd38 tempest-AttachInterfacesUnderV243Test-422957012 tempest-AttachInterfacesUnderV243Test-422957012-project-member] Lock "cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.293s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.568763] env[61987]: DEBUG nova.network.neutron [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Successfully updated port: b3159788-bb87-42ad-b6e5-201ff2afe647 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 998.624535] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062131, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.653908] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 998.654263] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-50719306-bdb4-4a6a-80a5-70eee357323e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.661316] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 998.661316] env[61987]: value = "task-1062132" [ 998.661316] env[61987]: _type = "Task" [ 998.661316] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.670218] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062132, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.849861] env[61987]: INFO nova.compute.manager [None req-ab627140-eb2c-416b-a1ec-22e753a078e1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Detaching volume 87789bc2-c4d6-4c37-ba7b-6211d2bb6126 [ 998.885622] env[61987]: INFO nova.virt.block_device [None req-ab627140-eb2c-416b-a1ec-22e753a078e1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Attempting to driver detach volume 87789bc2-c4d6-4c37-ba7b-6211d2bb6126 from mountpoint /dev/sdb [ 998.885882] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab627140-eb2c-416b-a1ec-22e753a078e1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Volume detach. Driver type: vmdk {{(pid=61987) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 998.886128] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab627140-eb2c-416b-a1ec-22e753a078e1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234411', 'volume_id': '87789bc2-c4d6-4c37-ba7b-6211d2bb6126', 'name': 'volume-87789bc2-c4d6-4c37-ba7b-6211d2bb6126', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '865222ae-12da-4e5a-84af-041003b50520', 'attached_at': '', 'detached_at': '', 'volume_id': '87789bc2-c4d6-4c37-ba7b-6211d2bb6126', 'serial': '87789bc2-c4d6-4c37-ba7b-6211d2bb6126'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 998.887061] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e09edf-88b0-492d-847f-b2a30605e4a1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.912099] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73d9abf4-5970-4849-8ee8-7221e016e4b1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.919450] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a48bc6f-fb68-40be-b7ac-7d2447556268 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.943181] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c8938ab-b88a-4520-a06d-1b1f34030db6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.960344] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab627140-eb2c-416b-a1ec-22e753a078e1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] The volume has not been displaced from its original location: [datastore2] volume-87789bc2-c4d6-4c37-ba7b-6211d2bb6126/volume-87789bc2-c4d6-4c37-ba7b-6211d2bb6126.vmdk. No consolidation needed. {{(pid=61987) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 998.966314] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab627140-eb2c-416b-a1ec-22e753a078e1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Reconfiguring VM instance instance-0000005e to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 998.966780] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5b09d5f-3093-4838-8216-46cb54d1ad9a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.986188] env[61987]: DEBUG oslo_vmware.api [None req-ab627140-eb2c-416b-a1ec-22e753a078e1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 998.986188] env[61987]: value = "task-1062133" [ 998.986188] env[61987]: _type = "Task" [ 998.986188] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.991894] env[61987]: INFO nova.compute.manager [None req-0334cc2a-9580-4a86-b414-c6d58fccaa2d tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Detaching volume fdf4dca1-67ae-4a18-911d-6151e71f109b [ 998.997911] env[61987]: DEBUG oslo_vmware.api [None req-ab627140-eb2c-416b-a1ec-22e753a078e1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062133, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.032261] env[61987]: INFO nova.virt.block_device [None req-0334cc2a-9580-4a86-b414-c6d58fccaa2d tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Attempting to driver detach volume fdf4dca1-67ae-4a18-911d-6151e71f109b from mountpoint /dev/sdb [ 999.032484] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0334cc2a-9580-4a86-b414-c6d58fccaa2d tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Volume detach. Driver type: vmdk {{(pid=61987) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 999.032805] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0334cc2a-9580-4a86-b414-c6d58fccaa2d tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234403', 'volume_id': 'fdf4dca1-67ae-4a18-911d-6151e71f109b', 'name': 'volume-fdf4dca1-67ae-4a18-911d-6151e71f109b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ae1396b3-c28b-4d1c-9a3a-3dc2170847ca', 'attached_at': '', 'detached_at': '', 'volume_id': 'fdf4dca1-67ae-4a18-911d-6151e71f109b', 'serial': 'fdf4dca1-67ae-4a18-911d-6151e71f109b'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 999.033732] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42cbe18e-082e-4495-938e-b89ebef4dc31 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.074304] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "refresh_cache-54d9bbcb-d01c-4e3b-8feb-4f8942d475d4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.074561] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired lock "refresh_cache-54d9bbcb-d01c-4e3b-8feb-4f8942d475d4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.074806] env[61987]: DEBUG nova.network.neutron [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 999.081029] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6630d75b-5f0a-4226-a9bb-7b656564aebf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.085611] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "4dc624ca-c23c-4c4f-8481-45f954e1c068" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.085989] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "4dc624ca-c23c-4c4f-8481-45f954e1c068" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.086303] env[61987]: DEBUG nova.compute.manager [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Going to confirm migration 4 {{(pid=61987) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5249}} [ 999.094720] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac441baf-9db5-4419-bd09-a47c4a5d4c08 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.124902] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f22054-a93e-4ed8-a164-74e48eb82188 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.133498] env[61987]: DEBUG oslo_vmware.api [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062131, 'name': PowerOnVM_Task, 'duration_secs': 0.604724} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.146098] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 999.148375] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0334cc2a-9580-4a86-b414-c6d58fccaa2d tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] The volume has not been displaced from its original location: [datastore2] volume-fdf4dca1-67ae-4a18-911d-6151e71f109b/volume-fdf4dca1-67ae-4a18-911d-6151e71f109b.vmdk. No consolidation needed. {{(pid=61987) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 999.153626] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0334cc2a-9580-4a86-b414-c6d58fccaa2d tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Reconfiguring VM instance instance-0000005a to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 999.155788] env[61987]: DEBUG nova.compute.manager [None req-3007d48e-7915-4319-bf14-defd368124ca tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 999.161016] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c8ef7e5-102d-40f6-a47e-31864786408a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.171642] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d24a0a1-e9af-45f5-a2fe-3072caa5d803 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.191132] env[61987]: DEBUG oslo_vmware.api [None req-0334cc2a-9580-4a86-b414-c6d58fccaa2d tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 999.191132] env[61987]: value = "task-1062134" [ 999.191132] env[61987]: _type = "Task" [ 999.191132] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.191399] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062132, 'name': PowerOffVM_Task, 'duration_secs': 0.38743} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.192440] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 999.196178] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f2e3888-e018-44f2-aa35-046b60e60322 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.205545] env[61987]: DEBUG oslo_vmware.api [None req-0334cc2a-9580-4a86-b414-c6d58fccaa2d tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062134, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.222749] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236e78ad-88fd-4137-a431-c9859608dbb1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.258482] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 999.258700] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7f042cb0-23d0-4df6-aed9-bea81ae09ac3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.268736] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 999.268736] env[61987]: value = "task-1062135" [ 999.268736] env[61987]: _type = "Task" [ 999.268736] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.279481] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] VM already powered off {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 999.279711] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 999.279960] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.280141] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.280336] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 999.280641] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d0471d22-e9d6-4a38-b860-3535e73f89c2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.289525] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 999.289726] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 999.293023] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bea0bcc3-bd86-4b3d-a698-d5258d9e45c2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.299311] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 999.299311] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]528ee54f-ff36-9eab-ba9f-cd65e3f22e18" [ 999.299311] env[61987]: _type = "Task" [ 999.299311] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.307150] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]528ee54f-ff36-9eab-ba9f-cd65e3f22e18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.321369] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24332d2e-30c1-45c0-badb-f604328c1ec1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.329317] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-130e54c8-299a-4531-9116-86cae64e36b6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.361013] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43641b49-f6bf-4843-8536-d434838f1cde {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.369387] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a040204d-4af2-4aca-b572-9b2209c48d89 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.383800] env[61987]: DEBUG nova.compute.provider_tree [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 999.497566] env[61987]: DEBUG oslo_vmware.api [None req-ab627140-eb2c-416b-a1ec-22e753a078e1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062133, 'name': ReconfigVM_Task, 'duration_secs': 0.371034} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.497900] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab627140-eb2c-416b-a1ec-22e753a078e1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Reconfigured VM instance instance-0000005e to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 999.505801] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27bc63ec-4d3c-4321-9a3a-9c71e91d107a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.521148] env[61987]: DEBUG oslo_vmware.api [None req-ab627140-eb2c-416b-a1ec-22e753a078e1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 999.521148] env[61987]: value = "task-1062136" [ 999.521148] env[61987]: _type = "Task" [ 999.521148] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.531353] env[61987]: DEBUG oslo_vmware.api [None req-ab627140-eb2c-416b-a1ec-22e753a078e1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062136, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.649567] env[61987]: DEBUG nova.network.neutron [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 999.707370] env[61987]: DEBUG oslo_vmware.api [None req-0334cc2a-9580-4a86-b414-c6d58fccaa2d tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062134, 'name': ReconfigVM_Task, 'duration_secs': 0.454422} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.707681] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0334cc2a-9580-4a86-b414-c6d58fccaa2d tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Reconfigured VM instance instance-0000005a to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 999.713219] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "refresh_cache-4dc624ca-c23c-4c4f-8481-45f954e1c068" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.713399] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquired lock "refresh_cache-4dc624ca-c23c-4c4f-8481-45f954e1c068" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.713577] env[61987]: DEBUG nova.network.neutron [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 999.713765] env[61987]: DEBUG nova.objects.instance [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lazy-loading 'info_cache' on Instance uuid 4dc624ca-c23c-4c4f-8481-45f954e1c068 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.715083] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-905b649e-0391-494e-904c-863f54244870 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.735628] env[61987]: DEBUG oslo_vmware.api [None req-0334cc2a-9580-4a86-b414-c6d58fccaa2d tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 999.735628] env[61987]: value = "task-1062137" [ 999.735628] env[61987]: _type = "Task" [ 999.735628] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.743922] env[61987]: DEBUG oslo_vmware.api [None req-0334cc2a-9580-4a86-b414-c6d58fccaa2d tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062137, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.806555] env[61987]: DEBUG nova.network.neutron [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Updating instance_info_cache with network_info: [{"id": "b3159788-bb87-42ad-b6e5-201ff2afe647", "address": "fa:16:3e:18:50:f4", "network": {"id": "d5ed1e9e-9033-441e-bfcb-2ac643d7f03f", "bridge": "br-int", "label": "tempest-ServersTestJSON-8466993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e62741abfb4471bbc7e11ae7d407ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39ab9baf-90cd-4fe2-8d56-434f8210fc19", "external-id": "nsx-vlan-transportzone-713", "segmentation_id": 713, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3159788-bb", "ovs_interfaceid": "b3159788-bb87-42ad-b6e5-201ff2afe647", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.815463] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]528ee54f-ff36-9eab-ba9f-cd65e3f22e18, 'name': SearchDatastore_Task, 'duration_secs': 0.018025} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.816364] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d69b789d-b314-44f1-a084-336e175f39cf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.822376] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 999.822376] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52505677-de59-3d35-8b54-9aac8d4728a9" [ 999.822376] env[61987]: _type = "Task" [ 999.822376] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.835224] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52505677-de59-3d35-8b54-9aac8d4728a9, 'name': SearchDatastore_Task, 'duration_secs': 0.01031} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.835495] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.835813] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] e6236bf1-4068-4a6e-9787-cd812381e122/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk. {{(pid=61987) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 999.836132] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-48e56fa1-4999-43d5-9fe9-57030eb20fc2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.844224] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 999.844224] env[61987]: value = "task-1062138" [ 999.844224] env[61987]: _type = "Task" [ 999.844224] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.852278] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062138, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.887214] env[61987]: DEBUG nova.scheduler.client.report [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1000.031853] env[61987]: DEBUG oslo_vmware.api [None req-ab627140-eb2c-416b-a1ec-22e753a078e1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062136, 'name': ReconfigVM_Task, 'duration_secs': 0.152295} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.032296] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab627140-eb2c-416b-a1ec-22e753a078e1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234411', 'volume_id': '87789bc2-c4d6-4c37-ba7b-6211d2bb6126', 'name': 'volume-87789bc2-c4d6-4c37-ba7b-6211d2bb6126', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '865222ae-12da-4e5a-84af-041003b50520', 'attached_at': '', 'detached_at': '', 'volume_id': '87789bc2-c4d6-4c37-ba7b-6211d2bb6126', 'serial': '87789bc2-c4d6-4c37-ba7b-6211d2bb6126'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1000.247482] env[61987]: DEBUG oslo_vmware.api [None req-0334cc2a-9580-4a86-b414-c6d58fccaa2d tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062137, 'name': ReconfigVM_Task, 'duration_secs': 0.204526} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.248523] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0334cc2a-9580-4a86-b414-c6d58fccaa2d tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234403', 'volume_id': 'fdf4dca1-67ae-4a18-911d-6151e71f109b', 'name': 'volume-fdf4dca1-67ae-4a18-911d-6151e71f109b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ae1396b3-c28b-4d1c-9a3a-3dc2170847ca', 'attached_at': '', 'detached_at': '', 'volume_id': 'fdf4dca1-67ae-4a18-911d-6151e71f109b', 'serial': 'fdf4dca1-67ae-4a18-911d-6151e71f109b'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1000.310141] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Releasing lock "refresh_cache-54d9bbcb-d01c-4e3b-8feb-4f8942d475d4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.310141] env[61987]: DEBUG nova.compute.manager [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Instance network_info: |[{"id": "b3159788-bb87-42ad-b6e5-201ff2afe647", "address": "fa:16:3e:18:50:f4", "network": {"id": "d5ed1e9e-9033-441e-bfcb-2ac643d7f03f", "bridge": "br-int", "label": "tempest-ServersTestJSON-8466993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e62741abfb4471bbc7e11ae7d407ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39ab9baf-90cd-4fe2-8d56-434f8210fc19", "external-id": "nsx-vlan-transportzone-713", "segmentation_id": 713, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3159788-bb", "ovs_interfaceid": "b3159788-bb87-42ad-b6e5-201ff2afe647", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1000.311214] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:18:50:f4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '39ab9baf-90cd-4fe2-8d56-434f8210fc19', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b3159788-bb87-42ad-b6e5-201ff2afe647', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1000.318917] env[61987]: DEBUG oslo.service.loopingcall [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1000.319884] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1000.319884] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9a3262a0-bee9-4ac9-9510-4744db218874 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.343824] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1000.343824] env[61987]: value = "task-1062139" [ 1000.343824] env[61987]: _type = "Task" [ 1000.343824] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.363017] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062138, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.363238] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062139, 'name': CreateVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.392081] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.392892] env[61987]: DEBUG nova.compute.manager [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1000.548226] env[61987]: DEBUG nova.compute.manager [req-7c18aa08-09a1-4ba9-b54a-6dafca766fea req-5d4fc74b-274a-4313-921a-997f35dd60cc service nova] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Received event network-changed-b3159788-bb87-42ad-b6e5-201ff2afe647 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1000.548475] env[61987]: DEBUG nova.compute.manager [req-7c18aa08-09a1-4ba9-b54a-6dafca766fea req-5d4fc74b-274a-4313-921a-997f35dd60cc service nova] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Refreshing instance network info cache due to event network-changed-b3159788-bb87-42ad-b6e5-201ff2afe647. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1000.548673] env[61987]: DEBUG oslo_concurrency.lockutils [req-7c18aa08-09a1-4ba9-b54a-6dafca766fea req-5d4fc74b-274a-4313-921a-997f35dd60cc service nova] Acquiring lock "refresh_cache-54d9bbcb-d01c-4e3b-8feb-4f8942d475d4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.548796] env[61987]: DEBUG oslo_concurrency.lockutils [req-7c18aa08-09a1-4ba9-b54a-6dafca766fea req-5d4fc74b-274a-4313-921a-997f35dd60cc service nova] Acquired lock "refresh_cache-54d9bbcb-d01c-4e3b-8feb-4f8942d475d4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.548976] env[61987]: DEBUG nova.network.neutron [req-7c18aa08-09a1-4ba9-b54a-6dafca766fea req-5d4fc74b-274a-4313-921a-997f35dd60cc service nova] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Refreshing network info cache for port b3159788-bb87-42ad-b6e5-201ff2afe647 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1000.600128] env[61987]: DEBUG nova.objects.instance [None req-ab627140-eb2c-416b-a1ec-22e753a078e1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lazy-loading 'flavor' on Instance uuid 865222ae-12da-4e5a-84af-041003b50520 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.807114] env[61987]: DEBUG nova.objects.instance [None req-0334cc2a-9580-4a86-b414-c6d58fccaa2d tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lazy-loading 'flavor' on Instance uuid ae1396b3-c28b-4d1c-9a3a-3dc2170847ca {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.853989] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062139, 'name': CreateVM_Task, 'duration_secs': 0.328674} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.856931] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1000.857993] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.858264] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.858610] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1000.859249] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19902118-0563-48d6-84d0-8d8d66c2678c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.864838] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062138, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51861} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.867483] env[61987]: INFO nova.virt.vmwareapi.ds_util [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] e6236bf1-4068-4a6e-9787-cd812381e122/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk. [ 1000.868267] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e002b5c-e7ca-40a9-9ece-04b44a476104 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.871867] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 1000.871867] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]526372d5-7095-6b8e-9090-0e412bc3324c" [ 1000.871867] env[61987]: _type = "Task" [ 1000.871867] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.897778] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] e6236bf1-4068-4a6e-9787-cd812381e122/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1000.902070] env[61987]: DEBUG nova.compute.utils [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1000.903364] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5858051-7428-4793-a0f0-831d82383873 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.921291] env[61987]: DEBUG nova.compute.manager [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1000.921468] env[61987]: DEBUG nova.network.neutron [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1000.923449] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]526372d5-7095-6b8e-9090-0e412bc3324c, 'name': SearchDatastore_Task, 'duration_secs': 0.011776} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.924265] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.924496] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1000.924722] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.924939] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.925114] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1000.925604] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dad07851-dcdc-4a95-9da0-bb234ddea711 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.928935] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1000.928935] env[61987]: value = "task-1062140" [ 1000.928935] env[61987]: _type = "Task" [ 1000.928935] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.935210] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1000.935384] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1000.939307] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-daad2c6c-f191-4fb1-b44e-8b20e8ea949c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.941782] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062140, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.945179] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 1000.945179] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]523476e1-e84b-009f-352a-ad5bcb955108" [ 1000.945179] env[61987]: _type = "Task" [ 1000.945179] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.954224] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523476e1-e84b-009f-352a-ad5bcb955108, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.986750] env[61987]: DEBUG nova.policy [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '697f9b1e3b3840369549a4f1ae3173ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd2afdc9bddba4ccab2b8fdb1f31da2b6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 1001.054233] env[61987]: DEBUG nova.network.neutron [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Updating instance_info_cache with network_info: [{"id": "1735c272-060e-4a5d-836a-af4e2eed480a", "address": "fa:16:3e:69:d8:56", "network": {"id": "ca68bf96-3f18-45ee-80fb-64a1260ed996", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-733595377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bd47eb44bc334bf3ae5813905903ecbc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a27fd90b-16a5-43af-bede-ae36762ece00", "external-id": "nsx-vlan-transportzone-197", "segmentation_id": 197, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1735c272-06", "ovs_interfaceid": "1735c272-060e-4a5d-836a-af4e2eed480a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.315284] env[61987]: DEBUG nova.network.neutron [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Successfully created port: fe549122-49ad-4628-b106-216d787be2d4 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1001.422011] env[61987]: DEBUG nova.compute.manager [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1001.446375] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062140, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.458598] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523476e1-e84b-009f-352a-ad5bcb955108, 'name': SearchDatastore_Task, 'duration_secs': 0.012091} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.462840] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a47a0650-98fa-42c7-944c-a2076005c1ed {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.469947] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 1001.469947] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5219bc7a-3aff-6945-2188-98d3fecaf587" [ 1001.469947] env[61987]: _type = "Task" [ 1001.469947] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.477954] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5219bc7a-3aff-6945-2188-98d3fecaf587, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.544601] env[61987]: DEBUG nova.network.neutron [req-7c18aa08-09a1-4ba9-b54a-6dafca766fea req-5d4fc74b-274a-4313-921a-997f35dd60cc service nova] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Updated VIF entry in instance network info cache for port b3159788-bb87-42ad-b6e5-201ff2afe647. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1001.545112] env[61987]: DEBUG nova.network.neutron [req-7c18aa08-09a1-4ba9-b54a-6dafca766fea req-5d4fc74b-274a-4313-921a-997f35dd60cc service nova] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Updating instance_info_cache with network_info: [{"id": "b3159788-bb87-42ad-b6e5-201ff2afe647", "address": "fa:16:3e:18:50:f4", "network": {"id": "d5ed1e9e-9033-441e-bfcb-2ac643d7f03f", "bridge": "br-int", "label": "tempest-ServersTestJSON-8466993-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "22e62741abfb4471bbc7e11ae7d407ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39ab9baf-90cd-4fe2-8d56-434f8210fc19", "external-id": "nsx-vlan-transportzone-713", "segmentation_id": 713, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3159788-bb", "ovs_interfaceid": "b3159788-bb87-42ad-b6e5-201ff2afe647", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.557383] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Releasing lock "refresh_cache-4dc624ca-c23c-4c4f-8481-45f954e1c068" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.557681] env[61987]: DEBUG nova.objects.instance [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lazy-loading 'migration_context' on Instance uuid 4dc624ca-c23c-4c4f-8481-45f954e1c068 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.610751] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ab627140-eb2c-416b-a1ec-22e753a078e1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "865222ae-12da-4e5a-84af-041003b50520" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.265s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.820676] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0334cc2a-9580-4a86-b414-c6d58fccaa2d tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.332s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.941223] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062140, 'name': ReconfigVM_Task, 'duration_secs': 0.888522} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.941495] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Reconfigured VM instance instance-00000065 to attach disk [datastore2] e6236bf1-4068-4a6e-9787-cd812381e122/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1001.942426] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b41236b-f691-4532-8704-14a69d379ba5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.969266] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c97b2aea-1abc-4c48-bf63-ad6b989c3179 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.996884] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5219bc7a-3aff-6945-2188-98d3fecaf587, 'name': SearchDatastore_Task, 'duration_secs': 0.020234} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.998441] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.998712] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4/54d9bbcb-d01c-4e3b-8feb-4f8942d475d4.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1001.999285] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1001.999285] env[61987]: value = "task-1062141" [ 1001.999285] env[61987]: _type = "Task" [ 1001.999285] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.999487] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eb2a3ad5-3871-4cc2-a59b-f7b4df2beb0e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.009644] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062141, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.010810] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 1002.010810] env[61987]: value = "task-1062142" [ 1002.010810] env[61987]: _type = "Task" [ 1002.010810] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.019433] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062142, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.048014] env[61987]: DEBUG oslo_concurrency.lockutils [req-7c18aa08-09a1-4ba9-b54a-6dafca766fea req-5d4fc74b-274a-4313-921a-997f35dd60cc service nova] Releasing lock "refresh_cache-54d9bbcb-d01c-4e3b-8feb-4f8942d475d4" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.061976] env[61987]: DEBUG nova.objects.base [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Object Instance<4dc624ca-c23c-4c4f-8481-45f954e1c068> lazy-loaded attributes: info_cache,migration_context {{(pid=61987) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1002.063012] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f04bbc48-2535-4c86-9319-5b65d0cde4ce {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.089395] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0975af2c-df49-4b6d-97c7-1a9bb90a313e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.095681] env[61987]: DEBUG oslo_vmware.api [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 1002.095681] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]529c03a0-074e-8d20-7410-27eede55a72f" [ 1002.095681] env[61987]: _type = "Task" [ 1002.095681] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.103288] env[61987]: DEBUG oslo_vmware.api [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]529c03a0-074e-8d20-7410-27eede55a72f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.423049] env[61987]: DEBUG oslo_concurrency.lockutils [None req-85c9f981-6c4a-416b-af21-c2ab2a8c4d90 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.423475] env[61987]: DEBUG oslo_concurrency.lockutils [None req-85c9f981-6c4a-416b-af21-c2ab2a8c4d90 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.429395] env[61987]: DEBUG oslo_concurrency.lockutils [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Acquiring lock "0da22c35-f11d-4498-bdb4-31bfd4c364e6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.429762] env[61987]: DEBUG oslo_concurrency.lockutils [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Lock "0da22c35-f11d-4498-bdb4-31bfd4c364e6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.431510] env[61987]: DEBUG nova.compute.manager [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1002.465125] env[61987]: DEBUG nova.virt.hardware [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1002.465509] env[61987]: DEBUG nova.virt.hardware [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1002.465696] env[61987]: DEBUG nova.virt.hardware [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1002.465963] env[61987]: DEBUG nova.virt.hardware [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1002.466157] env[61987]: DEBUG nova.virt.hardware [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1002.466318] env[61987]: DEBUG nova.virt.hardware [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1002.466636] env[61987]: DEBUG nova.virt.hardware [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1002.466893] env[61987]: DEBUG nova.virt.hardware [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1002.467121] env[61987]: DEBUG nova.virt.hardware [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1002.467340] env[61987]: DEBUG nova.virt.hardware [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1002.467564] env[61987]: DEBUG nova.virt.hardware [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1002.468717] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f971f8-c141-4df3-ab3a-6a22ac8120fa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.477670] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b323a236-7c66-41f3-a28c-c9e5c1374a78 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.493463] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "865222ae-12da-4e5a-84af-041003b50520" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.493743] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "865222ae-12da-4e5a-84af-041003b50520" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.493975] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "865222ae-12da-4e5a-84af-041003b50520-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.494186] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "865222ae-12da-4e5a-84af-041003b50520-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.494365] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "865222ae-12da-4e5a-84af-041003b50520-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.496861] env[61987]: INFO nova.compute.manager [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Terminating instance [ 1002.511174] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062141, 'name': ReconfigVM_Task, 'duration_secs': 0.203477} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.511546] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1002.514764] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4b555d80-feb8-45cf-a5bb-367189c31892 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.522022] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062142, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.523472] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1002.523472] env[61987]: value = "task-1062143" [ 1002.523472] env[61987]: _type = "Task" [ 1002.523472] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.531914] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062143, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.605901] env[61987]: DEBUG oslo_vmware.api [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]529c03a0-074e-8d20-7410-27eede55a72f, 'name': SearchDatastore_Task, 'duration_secs': 0.019048} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.606694] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.606694] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.719712] env[61987]: INFO nova.compute.manager [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Rescuing [ 1002.719996] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "refresh_cache-aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.720182] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquired lock "refresh_cache-aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.720360] env[61987]: DEBUG nova.network.neutron [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1002.768586] env[61987]: DEBUG nova.compute.manager [req-fd46029e-871a-4590-a3e6-46d6539336d5 req-de8bcc20-32f5-49d8-82c6-4a98f48bd9da service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Received event network-vif-plugged-fe549122-49ad-4628-b106-216d787be2d4 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1002.768861] env[61987]: DEBUG oslo_concurrency.lockutils [req-fd46029e-871a-4590-a3e6-46d6539336d5 req-de8bcc20-32f5-49d8-82c6-4a98f48bd9da service nova] Acquiring lock "7be2179d-68f7-4103-a8c1-b6e5b88b8706-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.769108] env[61987]: DEBUG oslo_concurrency.lockutils [req-fd46029e-871a-4590-a3e6-46d6539336d5 req-de8bcc20-32f5-49d8-82c6-4a98f48bd9da service nova] Lock "7be2179d-68f7-4103-a8c1-b6e5b88b8706-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.769330] env[61987]: DEBUG oslo_concurrency.lockutils [req-fd46029e-871a-4590-a3e6-46d6539336d5 req-de8bcc20-32f5-49d8-82c6-4a98f48bd9da service nova] Lock "7be2179d-68f7-4103-a8c1-b6e5b88b8706-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.769500] env[61987]: DEBUG nova.compute.manager [req-fd46029e-871a-4590-a3e6-46d6539336d5 req-de8bcc20-32f5-49d8-82c6-4a98f48bd9da service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] No waiting events found dispatching network-vif-plugged-fe549122-49ad-4628-b106-216d787be2d4 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1002.769674] env[61987]: WARNING nova.compute.manager [req-fd46029e-871a-4590-a3e6-46d6539336d5 req-de8bcc20-32f5-49d8-82c6-4a98f48bd9da service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Received unexpected event network-vif-plugged-fe549122-49ad-4628-b106-216d787be2d4 for instance with vm_state building and task_state spawning. [ 1002.839460] env[61987]: DEBUG nova.network.neutron [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Successfully updated port: fe549122-49ad-4628-b106-216d787be2d4 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1002.930524] env[61987]: INFO nova.compute.manager [None req-85c9f981-6c4a-416b-af21-c2ab2a8c4d90 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Detaching volume 9d3822d3-6b17-4fad-84da-5f273eb187c1 [ 1002.939626] env[61987]: DEBUG nova.compute.manager [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1002.968445] env[61987]: INFO nova.virt.block_device [None req-85c9f981-6c4a-416b-af21-c2ab2a8c4d90 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Attempting to driver detach volume 9d3822d3-6b17-4fad-84da-5f273eb187c1 from mountpoint /dev/sdc [ 1002.968727] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-85c9f981-6c4a-416b-af21-c2ab2a8c4d90 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Volume detach. Driver type: vmdk {{(pid=61987) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1002.968927] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-85c9f981-6c4a-416b-af21-c2ab2a8c4d90 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234412', 'volume_id': '9d3822d3-6b17-4fad-84da-5f273eb187c1', 'name': 'volume-9d3822d3-6b17-4fad-84da-5f273eb187c1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ae1396b3-c28b-4d1c-9a3a-3dc2170847ca', 'attached_at': '', 'detached_at': '', 'volume_id': '9d3822d3-6b17-4fad-84da-5f273eb187c1', 'serial': '9d3822d3-6b17-4fad-84da-5f273eb187c1'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1002.969883] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69a71ade-b851-4e06-ae00-b9bd4c3f3f1a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.995031] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf8ddbb1-2fd3-44f6-be3d-48198db3eb43 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.001900] env[61987]: DEBUG nova.compute.manager [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1003.002135] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1003.003228] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fee99daf-22fb-41e5-9617-a0e45fc506cb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.006166] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1429561c-75a0-4577-8a89-9af6e6c0f8c6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.012834] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1003.031255] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-354387f7-fffe-45f0-97f9-576eab436a3f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.036685] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a59e69c-1b4c-42e9-90a9-7dbdbdc4b53e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.043988] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062142, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.582462} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.059473] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4/54d9bbcb-d01c-4e3b-8feb-4f8942d475d4.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1003.059841] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1003.060738] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-85c9f981-6c4a-416b-af21-c2ab2a8c4d90 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] The volume has not been displaced from its original location: [datastore2] volume-9d3822d3-6b17-4fad-84da-5f273eb187c1/volume-9d3822d3-6b17-4fad-84da-5f273eb187c1.vmdk. No consolidation needed. {{(pid=61987) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1003.069478] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-85c9f981-6c4a-416b-af21-c2ab2a8c4d90 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Reconfiguring VM instance instance-0000005a to detach disk 2002 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1003.070041] env[61987]: DEBUG oslo_vmware.api [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 1003.070041] env[61987]: value = "task-1062144" [ 1003.070041] env[61987]: _type = "Task" [ 1003.070041] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.070352] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062143, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.070719] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e5915cb5-f9ea-40b9-a1fd-932983a043cb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.073102] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-deb045d8-802e-46fa-a7e4-19701b8f89f1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.094518] env[61987]: DEBUG oslo_vmware.api [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062144, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.096773] env[61987]: DEBUG oslo_vmware.api [None req-85c9f981-6c4a-416b-af21-c2ab2a8c4d90 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 1003.096773] env[61987]: value = "task-1062146" [ 1003.096773] env[61987]: _type = "Task" [ 1003.096773] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.097033] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 1003.097033] env[61987]: value = "task-1062145" [ 1003.097033] env[61987]: _type = "Task" [ 1003.097033] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.114175] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062145, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.117794] env[61987]: DEBUG oslo_vmware.api [None req-85c9f981-6c4a-416b-af21-c2ab2a8c4d90 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062146, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.313430] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8156db59-605c-4eb4-8eb4-ec5d3fc40667 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.324204] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b6516bd-f34a-45ac-a2c0-be319882c9fd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.355189] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.355356] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.355513] env[61987]: DEBUG nova.network.neutron [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1003.360013] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e658227-12ad-4ccb-9f6f-3ac747d25b4e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.368862] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e3b32f-45c2-4b35-9ac4-776bb0a92ff2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.386060] env[61987]: DEBUG nova.compute.provider_tree [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1003.460955] env[61987]: DEBUG oslo_concurrency.lockutils [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.468171] env[61987]: DEBUG nova.network.neutron [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Updating instance_info_cache with network_info: [{"id": "70bf0473-599e-4d1e-be30-efec5a322e37", "address": "fa:16:3e:5e:ec:e5", "network": {"id": "4ed4fbdc-05b0-44f8-a3d5-b7b288abd5e0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2067313775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a234a62da0e498fbe613fbcaaec3201", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70bf0473-59", "ovs_interfaceid": "70bf0473-599e-4d1e-be30-efec5a322e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.542925] env[61987]: DEBUG oslo_vmware.api [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062143, 'name': PowerOnVM_Task, 'duration_secs': 0.862784} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.543269] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1003.546093] env[61987]: DEBUG nova.compute.manager [None req-f0e2b60f-c871-4c65-97d8-cae51eb05055 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1003.546867] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e08496b7-9c4d-4ba8-b856-1cf9976830ca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.597937] env[61987]: DEBUG oslo_vmware.api [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062144, 'name': PowerOffVM_Task, 'duration_secs': 0.1988} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.603887] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1003.604171] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1003.604623] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-93aec9eb-f7cc-45e8-872e-c9df12a543c4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.611712] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062145, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.106983} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.614364] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1003.615034] env[61987]: DEBUG oslo_vmware.api [None req-85c9f981-6c4a-416b-af21-c2ab2a8c4d90 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062146, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.615508] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e075128f-5d59-4df1-819c-4fd6fb36bdb8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.639771] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4/54d9bbcb-d01c-4e3b-8feb-4f8942d475d4.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1003.640107] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a65e296-3314-475d-9fe3-814a38786f6d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.660590] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 1003.660590] env[61987]: value = "task-1062148" [ 1003.660590] env[61987]: _type = "Task" [ 1003.660590] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.671306] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062148, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.675921] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1003.676167] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1003.676389] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Deleting the datastore file [datastore1] 865222ae-12da-4e5a-84af-041003b50520 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1003.676627] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-089baf1f-ab2a-48a7-b371-ad6d1ae54d98 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.682912] env[61987]: DEBUG oslo_vmware.api [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 1003.682912] env[61987]: value = "task-1062149" [ 1003.682912] env[61987]: _type = "Task" [ 1003.682912] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.692815] env[61987]: DEBUG oslo_vmware.api [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062149, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.886600] env[61987]: DEBUG nova.network.neutron [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1003.889114] env[61987]: DEBUG nova.scheduler.client.report [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1003.971315] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Releasing lock "refresh_cache-aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.027785] env[61987]: DEBUG nova.network.neutron [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Updating instance_info_cache with network_info: [{"id": "fe549122-49ad-4628-b106-216d787be2d4", "address": "fa:16:3e:76:0c:4d", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe549122-49", "ovs_interfaceid": "fe549122-49ad-4628-b106-216d787be2d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.111799] env[61987]: DEBUG oslo_vmware.api [None req-85c9f981-6c4a-416b-af21-c2ab2a8c4d90 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062146, 'name': ReconfigVM_Task, 'duration_secs': 0.527427} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.112296] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-85c9f981-6c4a-416b-af21-c2ab2a8c4d90 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Reconfigured VM instance instance-0000005a to detach disk 2002 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1004.120433] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-44324fd1-3f24-473e-98c9-c8a4026c47b5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.143653] env[61987]: DEBUG oslo_vmware.api [None req-85c9f981-6c4a-416b-af21-c2ab2a8c4d90 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 1004.143653] env[61987]: value = "task-1062150" [ 1004.143653] env[61987]: _type = "Task" [ 1004.143653] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.156174] env[61987]: DEBUG oslo_vmware.api [None req-85c9f981-6c4a-416b-af21-c2ab2a8c4d90 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062150, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.169811] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062148, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.192947] env[61987]: DEBUG oslo_vmware.api [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062149, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.314359} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.192947] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1004.193153] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1004.193281] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1004.193500] env[61987]: INFO nova.compute.manager [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: 865222ae-12da-4e5a-84af-041003b50520] Took 1.19 seconds to destroy the instance on the hypervisor. [ 1004.193804] env[61987]: DEBUG oslo.service.loopingcall [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.194076] env[61987]: DEBUG nova.compute.manager [-] [instance: 865222ae-12da-4e5a-84af-041003b50520] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1004.194239] env[61987]: DEBUG nova.network.neutron [-] [instance: 865222ae-12da-4e5a-84af-041003b50520] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1004.530069] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.530418] env[61987]: DEBUG nova.compute.manager [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Instance network_info: |[{"id": "fe549122-49ad-4628-b106-216d787be2d4", "address": "fa:16:3e:76:0c:4d", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe549122-49", "ovs_interfaceid": "fe549122-49ad-4628-b106-216d787be2d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1004.530858] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:0c:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1c8fdf9-970c-4ae0-b6d9-f1015196b552', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fe549122-49ad-4628-b106-216d787be2d4', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1004.538758] env[61987]: DEBUG oslo.service.loopingcall [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.539313] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1004.539571] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b6b6b65-7133-4cca-97d5-95f2fb36228c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.559585] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1004.559585] env[61987]: value = "task-1062151" [ 1004.559585] env[61987]: _type = "Task" [ 1004.559585] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.568448] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062151, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.655226] env[61987]: DEBUG oslo_vmware.api [None req-85c9f981-6c4a-416b-af21-c2ab2a8c4d90 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062150, 'name': ReconfigVM_Task, 'duration_secs': 0.230034} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.655635] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-85c9f981-6c4a-416b-af21-c2ab2a8c4d90 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234412', 'volume_id': '9d3822d3-6b17-4fad-84da-5f273eb187c1', 'name': 'volume-9d3822d3-6b17-4fad-84da-5f273eb187c1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ae1396b3-c28b-4d1c-9a3a-3dc2170847ca', 'attached_at': '', 'detached_at': '', 'volume_id': '9d3822d3-6b17-4fad-84da-5f273eb187c1', 'serial': '9d3822d3-6b17-4fad-84da-5f273eb187c1'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1004.670989] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062148, 'name': ReconfigVM_Task, 'duration_secs': 0.553042} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.671287] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4/54d9bbcb-d01c-4e3b-8feb-4f8942d475d4.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1004.671913] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c601b364-d453-4b39-a727-e7e1aec72fe5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.679019] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 1004.679019] env[61987]: value = "task-1062152" [ 1004.679019] env[61987]: _type = "Task" [ 1004.679019] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.687723] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062152, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.815172] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquiring lock "dab6d2c5-8c04-4977-a409-b384037eabce" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.815449] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lock "dab6d2c5-8c04-4977-a409-b384037eabce" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.834857] env[61987]: DEBUG nova.compute.manager [req-63aaf5bf-93db-44aa-88ca-7646cf854ef2 req-bb60d892-cff4-4e23-b940-7072c4e52c11 service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Received event network-changed-fe549122-49ad-4628-b106-216d787be2d4 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1004.835158] env[61987]: DEBUG nova.compute.manager [req-63aaf5bf-93db-44aa-88ca-7646cf854ef2 req-bb60d892-cff4-4e23-b940-7072c4e52c11 service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Refreshing instance network info cache due to event network-changed-fe549122-49ad-4628-b106-216d787be2d4. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1004.835517] env[61987]: DEBUG oslo_concurrency.lockutils [req-63aaf5bf-93db-44aa-88ca-7646cf854ef2 req-bb60d892-cff4-4e23-b940-7072c4e52c11 service nova] Acquiring lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.835768] env[61987]: DEBUG oslo_concurrency.lockutils [req-63aaf5bf-93db-44aa-88ca-7646cf854ef2 req-bb60d892-cff4-4e23-b940-7072c4e52c11 service nova] Acquired lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.835981] env[61987]: DEBUG nova.network.neutron [req-63aaf5bf-93db-44aa-88ca-7646cf854ef2 req-bb60d892-cff4-4e23-b940-7072c4e52c11 service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Refreshing network info cache for port fe549122-49ad-4628-b106-216d787be2d4 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1004.902675] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.296s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.905663] env[61987]: DEBUG oslo_concurrency.lockutils [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.445s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.907235] env[61987]: INFO nova.compute.claims [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1005.069599] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062151, 'name': CreateVM_Task, 'duration_secs': 0.371583} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.069772] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1005.070526] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.070700] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.071045] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1005.071316] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b82a69b6-e311-43fc-a85b-8c997668aa90 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.076072] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1005.076072] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52f84391-9822-e80d-e162-a505dbf2add6" [ 1005.076072] env[61987]: _type = "Task" [ 1005.076072] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.083771] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52f84391-9822-e80d-e162-a505dbf2add6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.141238] env[61987]: DEBUG nova.network.neutron [-] [instance: 865222ae-12da-4e5a-84af-041003b50520] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.190674] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062152, 'name': Rename_Task, 'duration_secs': 0.17379} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.190674] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1005.190829] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-90f8e179-d294-42ad-aaea-a441179b4e21 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.197811] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 1005.197811] env[61987]: value = "task-1062153" [ 1005.197811] env[61987]: _type = "Task" [ 1005.197811] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.205622] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062153, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.207420] env[61987]: DEBUG nova.objects.instance [None req-85c9f981-6c4a-416b-af21-c2ab2a8c4d90 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lazy-loading 'flavor' on Instance uuid ae1396b3-c28b-4d1c-9a3a-3dc2170847ca {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.319250] env[61987]: DEBUG nova.compute.manager [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1005.467100] env[61987]: INFO nova.scheduler.client.report [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Deleted allocation for migration a799cdd4-fc8b-40af-a2f0-b591c5463c7d [ 1005.503491] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1005.503828] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-753aa47c-d5fc-44d9-afef-9b94e9ad31e8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.511901] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1005.511901] env[61987]: value = "task-1062154" [ 1005.511901] env[61987]: _type = "Task" [ 1005.511901] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.520107] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062154, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.560761] env[61987]: DEBUG nova.network.neutron [req-63aaf5bf-93db-44aa-88ca-7646cf854ef2 req-bb60d892-cff4-4e23-b940-7072c4e52c11 service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Updated VIF entry in instance network info cache for port fe549122-49ad-4628-b106-216d787be2d4. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1005.560761] env[61987]: DEBUG nova.network.neutron [req-63aaf5bf-93db-44aa-88ca-7646cf854ef2 req-bb60d892-cff4-4e23-b940-7072c4e52c11 service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Updating instance_info_cache with network_info: [{"id": "fe549122-49ad-4628-b106-216d787be2d4", "address": "fa:16:3e:76:0c:4d", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe549122-49", "ovs_interfaceid": "fe549122-49ad-4628-b106-216d787be2d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.587363] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52f84391-9822-e80d-e162-a505dbf2add6, 'name': SearchDatastore_Task, 'duration_secs': 0.041484} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.587703] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.587945] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1005.588217] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.588371] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.588557] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1005.588823] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c91342fb-89ed-4f32-96ed-2f0178cbecaf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.597029] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1005.597266] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1005.598053] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-293df449-f4da-426e-9740-fb70b036bb85 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.604341] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1005.604341] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5272a3e9-40c1-a24f-683e-5709b0733e7c" [ 1005.604341] env[61987]: _type = "Task" [ 1005.604341] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.615191] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5272a3e9-40c1-a24f-683e-5709b0733e7c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.643859] env[61987]: INFO nova.compute.manager [-] [instance: 865222ae-12da-4e5a-84af-041003b50520] Took 1.45 seconds to deallocate network for instance. [ 1005.707346] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062153, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.838102] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.974202] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "4dc624ca-c23c-4c4f-8481-45f954e1c068" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.888s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.023696] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062154, 'name': PowerOffVM_Task, 'duration_secs': 0.171677} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.023924] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1006.024768] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7aa8553-2010-4fa9-89ea-3f2de9d8eb44 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.045800] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8efd126b-a286-4f7d-9678-6ae71928d1e9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.063922] env[61987]: DEBUG oslo_concurrency.lockutils [req-63aaf5bf-93db-44aa-88ca-7646cf854ef2 req-bb60d892-cff4-4e23-b940-7072c4e52c11 service nova] Releasing lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.064311] env[61987]: DEBUG nova.compute.manager [req-63aaf5bf-93db-44aa-88ca-7646cf854ef2 req-bb60d892-cff4-4e23-b940-7072c4e52c11 service nova] [instance: 865222ae-12da-4e5a-84af-041003b50520] Received event network-vif-deleted-9df062f4-8440-4694-9896-000aa92195e2 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1006.064398] env[61987]: INFO nova.compute.manager [req-63aaf5bf-93db-44aa-88ca-7646cf854ef2 req-bb60d892-cff4-4e23-b940-7072c4e52c11 service nova] [instance: 865222ae-12da-4e5a-84af-041003b50520] Neutron deleted interface 9df062f4-8440-4694-9896-000aa92195e2; detaching it from the instance and deleting it from the info cache [ 1006.064573] env[61987]: DEBUG nova.network.neutron [req-63aaf5bf-93db-44aa-88ca-7646cf854ef2 req-bb60d892-cff4-4e23-b940-7072c4e52c11 service nova] [instance: 865222ae-12da-4e5a-84af-041003b50520] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.079834] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1006.080133] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe502e53-b6a9-4e30-a691-ba76ac18c544 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.086786] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1006.086786] env[61987]: value = "task-1062155" [ 1006.086786] env[61987]: _type = "Task" [ 1006.086786] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.095364] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062155, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.116840] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5272a3e9-40c1-a24f-683e-5709b0733e7c, 'name': SearchDatastore_Task, 'duration_secs': 0.009596} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.117735] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcd756ea-fa9b-443a-9f47-d070ace1d210 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.121794] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29971ab3-6c87-4207-87c7-e9fc055e56ea {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.125557] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1006.125557] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52ddfe72-21e7-edfa-943d-1fcd4d012f71" [ 1006.125557] env[61987]: _type = "Task" [ 1006.125557] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.131529] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31493d0a-2d97-412d-8a04-02e0cbe58a6c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.137287] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52ddfe72-21e7-edfa-943d-1fcd4d012f71, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.164820] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.166000] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c77e8bc-c8cb-4a75-aa42-148d6902fe6e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.173838] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-565c073a-581e-46a0-9256-5f966bc23ca3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.187695] env[61987]: DEBUG nova.compute.provider_tree [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.207699] env[61987]: DEBUG oslo_vmware.api [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062153, 'name': PowerOnVM_Task, 'duration_secs': 0.777698} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.207971] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1006.208202] env[61987]: INFO nova.compute.manager [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Took 7.93 seconds to spawn the instance on the hypervisor. [ 1006.208389] env[61987]: DEBUG nova.compute.manager [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1006.209184] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baf63e11-17a9-4da0-bf54-7df7ae295dae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.216082] env[61987]: DEBUG oslo_concurrency.lockutils [None req-85c9f981-6c4a-416b-af21-c2ab2a8c4d90 tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.793s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.568009] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eae40260-18fc-478c-902b-27e5b1608b46 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.577275] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcac46c6-1076-48b5-af0f-71d745eb7d15 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.599239] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] VM already powered off {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1006.599540] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1006.599758] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.599936] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.600159] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1006.613119] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4b823dbb-556b-4528-b29b-5c873bce57a9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.615925] env[61987]: DEBUG nova.compute.manager [req-63aaf5bf-93db-44aa-88ca-7646cf854ef2 req-bb60d892-cff4-4e23-b940-7072c4e52c11 service nova] [instance: 865222ae-12da-4e5a-84af-041003b50520] Detach interface failed, port_id=9df062f4-8440-4694-9896-000aa92195e2, reason: Instance 865222ae-12da-4e5a-84af-041003b50520 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 1006.616821] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "4dc624ca-c23c-4c4f-8481-45f954e1c068" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.616926] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "4dc624ca-c23c-4c4f-8481-45f954e1c068" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.617307] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "4dc624ca-c23c-4c4f-8481-45f954e1c068-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.617523] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "4dc624ca-c23c-4c4f-8481-45f954e1c068-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.617739] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "4dc624ca-c23c-4c4f-8481-45f954e1c068-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.619913] env[61987]: INFO nova.compute.manager [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Terminating instance [ 1006.627938] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1006.628152] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1006.632837] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b96791f9-ce45-4e4b-8c3d-5ba70bca0925 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.641614] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52ddfe72-21e7-edfa-943d-1fcd4d012f71, 'name': SearchDatastore_Task, 'duration_secs': 0.013194} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.642730] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.643022] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 7be2179d-68f7-4103-a8c1-b6e5b88b8706/7be2179d-68f7-4103-a8c1-b6e5b88b8706.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1006.643362] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1006.643362] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52f3095d-5e4b-28cb-55d9-d34dbbe03e82" [ 1006.643362] env[61987]: _type = "Task" [ 1006.643362] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.643571] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a040bea-33e1-403b-b85a-7cc06fad235c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.654862] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52f3095d-5e4b-28cb-55d9-d34dbbe03e82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.655977] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1006.655977] env[61987]: value = "task-1062156" [ 1006.655977] env[61987]: _type = "Task" [ 1006.655977] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.663169] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062156, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.691336] env[61987]: DEBUG nova.scheduler.client.report [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1006.729160] env[61987]: INFO nova.compute.manager [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Took 12.84 seconds to build instance. [ 1006.842857] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.843139] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.843365] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.843561] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.843740] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.845957] env[61987]: INFO nova.compute.manager [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Terminating instance [ 1007.125162] env[61987]: DEBUG nova.compute.manager [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1007.125385] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1007.126431] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-408548e7-ae68-453a-a0ed-8013bf96a94e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.135124] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1007.135409] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f607567-dee2-453b-a119-21d268c9bc4f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.142426] env[61987]: DEBUG oslo_vmware.api [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 1007.142426] env[61987]: value = "task-1062157" [ 1007.142426] env[61987]: _type = "Task" [ 1007.142426] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.153593] env[61987]: DEBUG oslo_vmware.api [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062157, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.158016] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52f3095d-5e4b-28cb-55d9-d34dbbe03e82, 'name': SearchDatastore_Task, 'duration_secs': 0.014653} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.161760] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcefb97d-cfe4-4f1f-ab50-bcd3bcebd4be {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.170334] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1007.170334] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52fb6582-bbe0-85be-3951-5e0e7116e920" [ 1007.170334] env[61987]: _type = "Task" [ 1007.170334] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.170610] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062156, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.181035] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52fb6582-bbe0-85be-3951-5e0e7116e920, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.196844] env[61987]: DEBUG oslo_concurrency.lockutils [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.291s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.197478] env[61987]: DEBUG nova.compute.manager [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1007.200532] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.363s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.202211] env[61987]: INFO nova.compute.claims [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1007.231717] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d281c49a-a678-4fe5-8e6a-1acbdb8c2584 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "54d9bbcb-d01c-4e3b-8feb-4f8942d475d4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.353s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.349586] env[61987]: DEBUG nova.compute.manager [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1007.349997] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1007.350965] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4aaa8e1-5898-4283-91b2-62e08f98bb5e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.358905] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1007.359178] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e088f186-5d2a-4459-9081-27ee47835922 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.366903] env[61987]: DEBUG oslo_vmware.api [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 1007.366903] env[61987]: value = "task-1062158" [ 1007.366903] env[61987]: _type = "Task" [ 1007.366903] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.375021] env[61987]: DEBUG oslo_vmware.api [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062158, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.652468] env[61987]: DEBUG oslo_vmware.api [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062157, 'name': PowerOffVM_Task, 'duration_secs': 0.273351} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.652743] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1007.652919] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1007.653180] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-07b8810b-19f4-4262-998b-67af9f0b4b7b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.665299] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062156, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.683576} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.665532] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 7be2179d-68f7-4103-a8c1-b6e5b88b8706/7be2179d-68f7-4103-a8c1-b6e5b88b8706.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1007.665747] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1007.665984] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5c2b2d3e-2fd4-44ce-a4ab-d47849226215 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.671815] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1007.671815] env[61987]: value = "task-1062160" [ 1007.671815] env[61987]: _type = "Task" [ 1007.671815] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.689535] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062160, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.689764] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52fb6582-bbe0-85be-3951-5e0e7116e920, 'name': SearchDatastore_Task, 'duration_secs': 0.019421} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.690565] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.690694] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk. {{(pid=61987) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1007.691453] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a8e16d4f-c369-45e0-9c84-7fedd1be14bb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.699613] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1007.699613] env[61987]: value = "task-1062161" [ 1007.699613] env[61987]: _type = "Task" [ 1007.699613] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.706714] env[61987]: DEBUG nova.compute.utils [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1007.712797] env[61987]: DEBUG nova.compute.manager [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Not allocating networking since 'none' was specified. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 1007.713357] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062161, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.734539] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1007.735036] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1007.735036] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Deleting the datastore file [datastore2] 4dc624ca-c23c-4c4f-8481-45f954e1c068 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1007.735268] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a31420cb-ece4-468c-9d18-388ac041cfd6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.744893] env[61987]: DEBUG oslo_vmware.api [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for the task: (returnval){ [ 1007.744893] env[61987]: value = "task-1062162" [ 1007.744893] env[61987]: _type = "Task" [ 1007.744893] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.753314] env[61987]: DEBUG oslo_vmware.api [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062162, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.877263] env[61987]: DEBUG oslo_vmware.api [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062158, 'name': PowerOffVM_Task, 'duration_secs': 0.247654} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.877555] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1007.877765] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1007.877988] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c79b7433-3468-4bbe-af5d-a548260569bf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.975412] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1007.975656] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1007.975850] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Deleting the datastore file [datastore2] ae1396b3-c28b-4d1c-9a3a-3dc2170847ca {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1007.976184] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aabbe200-1f1d-4043-bbb0-974e3d5b2439 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.983170] env[61987]: DEBUG oslo_vmware.api [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for the task: (returnval){ [ 1007.983170] env[61987]: value = "task-1062164" [ 1007.983170] env[61987]: _type = "Task" [ 1007.983170] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.993515] env[61987]: DEBUG oslo_vmware.api [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062164, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.185802] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062160, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073008} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.186136] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1008.187033] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad8f1849-554f-42b8-8619-5f1d5b8fe757 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.191461] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0cc4ad70-0b45-47b1-b4a6-d83661b4af73 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "54d9bbcb-d01c-4e3b-8feb-4f8942d475d4" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.191723] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0cc4ad70-0b45-47b1-b4a6-d83661b4af73 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "54d9bbcb-d01c-4e3b-8feb-4f8942d475d4" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.191900] env[61987]: DEBUG nova.compute.manager [None req-0cc4ad70-0b45-47b1-b4a6-d83661b4af73 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1008.202423] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4730269e-223e-4a9b-9b30-71c4a1d5fa74 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.214158] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] 7be2179d-68f7-4103-a8c1-b6e5b88b8706/7be2179d-68f7-4103-a8c1-b6e5b88b8706.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1008.218035] env[61987]: DEBUG nova.compute.manager [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1008.223012] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8807704-5c3e-479d-878c-8ae462d3a43e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.242572] env[61987]: DEBUG nova.compute.manager [None req-0cc4ad70-0b45-47b1-b4a6-d83661b4af73 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61987) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1008.243237] env[61987]: DEBUG nova.objects.instance [None req-0cc4ad70-0b45-47b1-b4a6-d83661b4af73 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lazy-loading 'flavor' on Instance uuid 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.248208] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062161, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.248516] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1008.248516] env[61987]: value = "task-1062165" [ 1008.248516] env[61987]: _type = "Task" [ 1008.248516] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.255737] env[61987]: DEBUG nova.scheduler.client.report [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Refreshing inventories for resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1008.264221] env[61987]: DEBUG oslo_vmware.api [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Task: {'id': task-1062162, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.35749} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.267867] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1008.268195] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1008.268387] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1008.268574] env[61987]: INFO nova.compute.manager [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1008.268831] env[61987]: DEBUG oslo.service.loopingcall [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1008.269402] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062165, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.269652] env[61987]: DEBUG nova.compute.manager [-] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1008.269763] env[61987]: DEBUG nova.network.neutron [-] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1008.273096] env[61987]: DEBUG nova.scheduler.client.report [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Updating ProviderTree inventory for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1008.273356] env[61987]: DEBUG nova.compute.provider_tree [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1008.286589] env[61987]: DEBUG nova.scheduler.client.report [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Refreshing aggregate associations for resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0, aggregates: None {{(pid=61987) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1008.305198] env[61987]: DEBUG nova.scheduler.client.report [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Refreshing trait associations for resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61987) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1008.495153] env[61987]: DEBUG oslo_vmware.api [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062164, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.497707] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e19bb8c-9235-4296-b8dd-ac4d94e7fdca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.507065] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43edc929-6e2b-40fb-810c-1a2c4f66ba4c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.541237] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bccb77d-6e05-4678-af02-528f4085dd5a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.548970] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9ba812-3c25-48d9-a9f7-5ca33a9a150a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.562962] env[61987]: DEBUG nova.compute.provider_tree [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1008.716137] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062161, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.571174} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.716418] env[61987]: INFO nova.virt.vmwareapi.ds_util [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk. [ 1008.717251] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b9a7ba7-feb0-4f76-97be-95d06ced6bf6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.745102] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1008.748727] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6005cf6-6971-4441-8d5d-1207c311ae44 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.774621] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062165, 'name': ReconfigVM_Task, 'duration_secs': 0.321064} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.776362] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Reconfigured VM instance instance-00000067 to attach disk [datastore1] 7be2179d-68f7-4103-a8c1-b6e5b88b8706/7be2179d-68f7-4103-a8c1-b6e5b88b8706.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1008.777079] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1008.777079] env[61987]: value = "task-1062166" [ 1008.777079] env[61987]: _type = "Task" [ 1008.777079] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.777534] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-42b50f7e-5925-4bb9-b9bb-74bc81061ec7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.788267] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062166, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.789679] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1008.789679] env[61987]: value = "task-1062167" [ 1008.789679] env[61987]: _type = "Task" [ 1008.789679] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.802445] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062167, 'name': Rename_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.887384] env[61987]: DEBUG nova.compute.manager [req-b399e417-b5da-44da-b626-90cefc928e42 req-1aeb4b57-5444-446a-9d8b-94b36fb911db service nova] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Received event network-vif-deleted-1735c272-060e-4a5d-836a-af4e2eed480a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1008.887627] env[61987]: INFO nova.compute.manager [req-b399e417-b5da-44da-b626-90cefc928e42 req-1aeb4b57-5444-446a-9d8b-94b36fb911db service nova] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Neutron deleted interface 1735c272-060e-4a5d-836a-af4e2eed480a; detaching it from the instance and deleting it from the info cache [ 1008.887884] env[61987]: DEBUG nova.network.neutron [req-b399e417-b5da-44da-b626-90cefc928e42 req-1aeb4b57-5444-446a-9d8b-94b36fb911db service nova] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.993529] env[61987]: DEBUG oslo_vmware.api [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Task: {'id': task-1062164, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.576895} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.993789] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1008.994046] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1008.995113] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1008.995113] env[61987]: INFO nova.compute.manager [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Took 1.64 seconds to destroy the instance on the hypervisor. [ 1008.995113] env[61987]: DEBUG oslo.service.loopingcall [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1008.995113] env[61987]: DEBUG nova.compute.manager [-] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1008.995113] env[61987]: DEBUG nova.network.neutron [-] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1009.068060] env[61987]: DEBUG nova.scheduler.client.report [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1009.263917] env[61987]: DEBUG nova.compute.manager [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1009.270778] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cc4ad70-0b45-47b1-b4a6-d83661b4af73 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1009.271088] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dab3ca07-011e-437b-9d3f-4796b71b3c6e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.278763] env[61987]: DEBUG oslo_vmware.api [None req-0cc4ad70-0b45-47b1-b4a6-d83661b4af73 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 1009.278763] env[61987]: value = "task-1062168" [ 1009.278763] env[61987]: _type = "Task" [ 1009.278763] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.291727] env[61987]: DEBUG oslo_vmware.api [None req-0cc4ad70-0b45-47b1-b4a6-d83661b4af73 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062168, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.299664] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062166, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.305050] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062167, 'name': Rename_Task, 'duration_secs': 0.153006} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.307289] env[61987]: DEBUG nova.virt.hardware [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1009.307663] env[61987]: DEBUG nova.virt.hardware [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1009.307876] env[61987]: DEBUG nova.virt.hardware [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1009.307876] env[61987]: DEBUG nova.virt.hardware [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1009.308096] env[61987]: DEBUG nova.virt.hardware [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1009.308194] env[61987]: DEBUG nova.virt.hardware [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1009.309219] env[61987]: DEBUG nova.virt.hardware [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1009.309219] env[61987]: DEBUG nova.virt.hardware [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1009.309219] env[61987]: DEBUG nova.virt.hardware [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1009.309219] env[61987]: DEBUG nova.virt.hardware [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1009.309219] env[61987]: DEBUG nova.virt.hardware [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1009.309420] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1009.310129] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4666d3a6-2336-4572-a915-2c503efa7ea3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.313133] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e2a78b35-c54c-4acd-9d86-a2fa38945fb0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.320518] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a85fe169-6f17-4bbf-9f6a-e5a858bb90e6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.324457] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1009.324457] env[61987]: value = "task-1062169" [ 1009.324457] env[61987]: _type = "Task" [ 1009.324457] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.336499] env[61987]: DEBUG nova.network.neutron [-] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.338727] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Instance VIF info [] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1009.344124] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Creating folder: Project (a039c39305fc463d9ff7de1dc3ec90e3). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1009.345244] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fafb8a7b-df06-468a-b0fc-95c1110a0e07 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.351330] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062169, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.358805] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Created folder: Project (a039c39305fc463d9ff7de1dc3ec90e3) in parent group-v234219. [ 1009.358846] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Creating folder: Instances. Parent ref: group-v234415. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1009.359087] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a21d39ae-fc06-473c-b414-ee5361466685 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.368635] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Created folder: Instances in parent group-v234415. [ 1009.368904] env[61987]: DEBUG oslo.service.loopingcall [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.369096] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1009.369314] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f76649fa-1d1d-4a71-b038-18b148c8b111 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.394373] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ae6cfc45-a322-4624-90bb-322eaed9ed2b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.395892] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1009.395892] env[61987]: value = "task-1062172" [ 1009.395892] env[61987]: _type = "Task" [ 1009.395892] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.404109] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f844997c-f444-4762-9326-30525d555a2b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.419193] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062172, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.438539] env[61987]: DEBUG nova.compute.manager [req-b399e417-b5da-44da-b626-90cefc928e42 req-1aeb4b57-5444-446a-9d8b-94b36fb911db service nova] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Detach interface failed, port_id=1735c272-060e-4a5d-836a-af4e2eed480a, reason: Instance 4dc624ca-c23c-4c4f-8481-45f954e1c068 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 1009.472158] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "f3773cda-41e8-4804-9319-c7e458112ddf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.472441] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "f3773cda-41e8-4804-9319-c7e458112ddf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.572055] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.371s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.572605] env[61987]: DEBUG nova.compute.manager [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1009.575191] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.410s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.575426] env[61987]: DEBUG nova.objects.instance [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lazy-loading 'resources' on Instance uuid 865222ae-12da-4e5a-84af-041003b50520 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1009.790289] env[61987]: DEBUG oslo_vmware.api [None req-0cc4ad70-0b45-47b1-b4a6-d83661b4af73 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062168, 'name': PowerOffVM_Task, 'duration_secs': 0.193236} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.791378] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cc4ad70-0b45-47b1-b4a6-d83661b4af73 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1009.791777] env[61987]: DEBUG nova.compute.manager [None req-0cc4ad70-0b45-47b1-b4a6-d83661b4af73 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1009.792769] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2727b1f-31b1-484c-a436-1bfe330ed758 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.799990] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062166, 'name': ReconfigVM_Task, 'duration_secs': 0.659512} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.800943] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Reconfigured VM instance instance-00000063 to attach disk [datastore2] aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1009.802033] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc2f0789-a4c9-4bed-a4fa-4490022abcba {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.845958] env[61987]: INFO nova.compute.manager [-] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Took 1.58 seconds to deallocate network for instance. [ 1009.846317] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7b90a8b-4fa7-4da1-9935-1d4c741d55e3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.868635] env[61987]: DEBUG oslo_vmware.api [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062169, 'name': PowerOnVM_Task, 'duration_secs': 0.523608} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.869181] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1009.869181] env[61987]: value = "task-1062173" [ 1009.869181] env[61987]: _type = "Task" [ 1009.869181] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.869408] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1009.870118] env[61987]: INFO nova.compute.manager [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Took 7.44 seconds to spawn the instance on the hypervisor. [ 1009.870118] env[61987]: DEBUG nova.compute.manager [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1009.870611] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37903059-beb6-42db-970a-e361808f26b0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.885328] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062173, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.905936] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062172, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.975989] env[61987]: DEBUG nova.compute.manager [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1010.021016] env[61987]: DEBUG nova.network.neutron [-] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.080136] env[61987]: DEBUG nova.compute.utils [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1010.086260] env[61987]: DEBUG nova.compute.manager [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1010.086260] env[61987]: DEBUG nova.network.neutron [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1010.130636] env[61987]: DEBUG nova.policy [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06accdea3e014922af1f011175c36bc6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '630439422a354de0b32ac5eefaa990b9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 1010.288678] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-590539f6-3b23-4e40-b3b2-83eaaae7f948 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.298020] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3868852c-a9c0-4926-8bf6-779994b29a85 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.336851] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c15cb0e3-a009-4a5b-ae50-dd4ccfcb2dd9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.341702] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0cc4ad70-0b45-47b1-b4a6-d83661b4af73 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "54d9bbcb-d01c-4e3b-8feb-4f8942d475d4" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.150s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.348293] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af180ab-b852-4c0e-863d-ab57a4f912b3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.365109] env[61987]: DEBUG nova.compute.provider_tree [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1010.367354] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.379793] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062173, 'name': ReconfigVM_Task, 'duration_secs': 0.152805} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.380130] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1010.380626] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c09e4cee-35b1-4fae-9197-568c90ae156f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.390278] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1010.390278] env[61987]: value = "task-1062174" [ 1010.390278] env[61987]: _type = "Task" [ 1010.390278] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.396011] env[61987]: INFO nova.compute.manager [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Took 13.37 seconds to build instance. [ 1010.406749] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062174, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.412806] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062172, 'name': CreateVM_Task, 'duration_secs': 0.533141} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.412904] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1010.413542] env[61987]: DEBUG oslo_concurrency.lockutils [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.413902] env[61987]: DEBUG oslo_concurrency.lockutils [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.414240] env[61987]: DEBUG oslo_concurrency.lockutils [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1010.414749] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f344cf60-d6ee-4181-bad7-9c24c4ae03e2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.419267] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for the task: (returnval){ [ 1010.419267] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5253a56b-7416-fddf-8fd9-7f7794d9ae3a" [ 1010.419267] env[61987]: _type = "Task" [ 1010.419267] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.426896] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5253a56b-7416-fddf-8fd9-7f7794d9ae3a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.497065] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.524468] env[61987]: INFO nova.compute.manager [-] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Took 1.53 seconds to deallocate network for instance. [ 1010.562536] env[61987]: DEBUG nova.network.neutron [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Successfully created port: 05cdb86d-67c4-4b34-9cd3-4e4b3ed04d58 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1010.586481] env[61987]: DEBUG nova.compute.manager [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1010.869491] env[61987]: DEBUG nova.scheduler.client.report [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1010.903281] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5c62ba97-a02d-4297-8cab-0d92b3e8638c tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "7be2179d-68f7-4103-a8c1-b6e5b88b8706" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.894s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.912012] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062174, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.930860] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5253a56b-7416-fddf-8fd9-7f7794d9ae3a, 'name': SearchDatastore_Task, 'duration_secs': 0.008587} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.931182] env[61987]: DEBUG oslo_concurrency.lockutils [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.931421] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1010.931800] env[61987]: DEBUG oslo_concurrency.lockutils [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.931995] env[61987]: DEBUG oslo_concurrency.lockutils [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.932251] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1010.933342] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a8cee3f8-95f6-41e8-97fe-5480c11c2699 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.941601] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1010.941954] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1010.942802] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c5d6642-ba1f-4091-992b-95855a8cc767 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.948393] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for the task: (returnval){ [ 1010.948393] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]524e5b31-50a9-1ba1-32a7-84051fb7b7af" [ 1010.948393] env[61987]: _type = "Task" [ 1010.948393] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.956695] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524e5b31-50a9-1ba1-32a7-84051fb7b7af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.035267] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.126331] env[61987]: DEBUG nova.compute.manager [req-70b561c6-d346-483e-bf66-7ccc57be6663 req-af3eda01-5dc2-4c06-8d51-e630d54865af service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Received event network-changed-fe549122-49ad-4628-b106-216d787be2d4 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1011.126331] env[61987]: DEBUG nova.compute.manager [req-70b561c6-d346-483e-bf66-7ccc57be6663 req-af3eda01-5dc2-4c06-8d51-e630d54865af service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Refreshing instance network info cache due to event network-changed-fe549122-49ad-4628-b106-216d787be2d4. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1011.126331] env[61987]: DEBUG oslo_concurrency.lockutils [req-70b561c6-d346-483e-bf66-7ccc57be6663 req-af3eda01-5dc2-4c06-8d51-e630d54865af service nova] Acquiring lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.126484] env[61987]: DEBUG oslo_concurrency.lockutils [req-70b561c6-d346-483e-bf66-7ccc57be6663 req-af3eda01-5dc2-4c06-8d51-e630d54865af service nova] Acquired lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.126666] env[61987]: DEBUG nova.network.neutron [req-70b561c6-d346-483e-bf66-7ccc57be6663 req-af3eda01-5dc2-4c06-8d51-e630d54865af service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Refreshing network info cache for port fe549122-49ad-4628-b106-216d787be2d4 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1011.277943] env[61987]: DEBUG nova.compute.manager [req-01d95e51-072a-4ad2-9e70-564e2c93672e req-4afec2d4-4bb6-49bb-824e-6f4e05736d35 service nova] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Received event network-vif-deleted-57c4615c-de74-4af1-8b67-1801365b0ea4 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1011.378533] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.803s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.381677] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.014s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.381677] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.383133] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.886s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.384625] env[61987]: INFO nova.compute.claims [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1011.398237] env[61987]: INFO nova.scheduler.client.report [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Deleted allocations for instance 865222ae-12da-4e5a-84af-041003b50520 [ 1011.407500] env[61987]: DEBUG oslo_vmware.api [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062174, 'name': PowerOnVM_Task, 'duration_secs': 0.98491} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.407500] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1011.408296] env[61987]: DEBUG nova.compute.manager [None req-0b400225-42bb-45a6-a475-57aa26012b5f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1011.410386] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30c99922-7830-42b1-87e1-6920c9e79017 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.417478] env[61987]: INFO nova.scheduler.client.report [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Deleted allocations for instance 4dc624ca-c23c-4c4f-8481-45f954e1c068 [ 1011.462359] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524e5b31-50a9-1ba1-32a7-84051fb7b7af, 'name': SearchDatastore_Task, 'duration_secs': 0.009824} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.463815] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dae5996a-7cb9-4f06-8f0d-f894c856c27a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.470181] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for the task: (returnval){ [ 1011.470181] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52a8f3b9-46e4-d6f7-4e75-e2e54ca4e947" [ 1011.470181] env[61987]: _type = "Task" [ 1011.470181] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.478098] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52a8f3b9-46e4-d6f7-4e75-e2e54ca4e947, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.596800] env[61987]: DEBUG nova.compute.manager [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1011.622665] env[61987]: DEBUG nova.virt.hardware [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1011.622903] env[61987]: DEBUG nova.virt.hardware [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1011.623129] env[61987]: DEBUG nova.virt.hardware [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1011.623296] env[61987]: DEBUG nova.virt.hardware [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1011.623581] env[61987]: DEBUG nova.virt.hardware [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1011.623792] env[61987]: DEBUG nova.virt.hardware [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1011.624067] env[61987]: DEBUG nova.virt.hardware [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1011.624254] env[61987]: DEBUG nova.virt.hardware [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1011.624439] env[61987]: DEBUG nova.virt.hardware [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1011.624610] env[61987]: DEBUG nova.virt.hardware [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1011.624791] env[61987]: DEBUG nova.virt.hardware [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1011.625728] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e04587b-28ef-48bf-b095-f4dc30dad755 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.636741] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f3c5197-d1b9-4f2d-a5eb-c17518881bc7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.738949] env[61987]: DEBUG oslo_concurrency.lockutils [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "b172195a-79d2-4f24-b9df-320a07de965e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.739211] env[61987]: DEBUG oslo_concurrency.lockutils [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "b172195a-79d2-4f24-b9df-320a07de965e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.758157] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "54d9bbcb-d01c-4e3b-8feb-4f8942d475d4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.758438] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "54d9bbcb-d01c-4e3b-8feb-4f8942d475d4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.758654] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "54d9bbcb-d01c-4e3b-8feb-4f8942d475d4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.758845] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "54d9bbcb-d01c-4e3b-8feb-4f8942d475d4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.759036] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "54d9bbcb-d01c-4e3b-8feb-4f8942d475d4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.761594] env[61987]: INFO nova.compute.manager [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Terminating instance [ 1011.802653] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "0cba5e51-7e0a-409b-a7b2-612ab8cf5689" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.802973] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "0cba5e51-7e0a-409b-a7b2-612ab8cf5689" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.859835] env[61987]: DEBUG nova.network.neutron [req-70b561c6-d346-483e-bf66-7ccc57be6663 req-af3eda01-5dc2-4c06-8d51-e630d54865af service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Updated VIF entry in instance network info cache for port fe549122-49ad-4628-b106-216d787be2d4. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1011.860243] env[61987]: DEBUG nova.network.neutron [req-70b561c6-d346-483e-bf66-7ccc57be6663 req-af3eda01-5dc2-4c06-8d51-e630d54865af service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Updating instance_info_cache with network_info: [{"id": "fe549122-49ad-4628-b106-216d787be2d4", "address": "fa:16:3e:76:0c:4d", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe549122-49", "ovs_interfaceid": "fe549122-49ad-4628-b106-216d787be2d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.907810] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5419be06-d55e-4d21-b0cd-16a5ec506f1a tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "865222ae-12da-4e5a-84af-041003b50520" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.414s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.935235] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a5ea086-83e8-45e7-9691-f98be2b48e68 tempest-DeleteServersTestJSON-1141004998 tempest-DeleteServersTestJSON-1141004998-project-member] Lock "4dc624ca-c23c-4c4f-8481-45f954e1c068" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.318s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.979892] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52a8f3b9-46e4-d6f7-4e75-e2e54ca4e947, 'name': SearchDatastore_Task, 'duration_secs': 0.020946} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.980245] env[61987]: DEBUG oslo_concurrency.lockutils [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.980550] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 0da22c35-f11d-4498-bdb4-31bfd4c364e6/0da22c35-f11d-4498-bdb4-31bfd4c364e6.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1011.980847] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ed275d82-7bc4-48f8-9d18-ef1395609cf2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.987189] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for the task: (returnval){ [ 1011.987189] env[61987]: value = "task-1062175" [ 1011.987189] env[61987]: _type = "Task" [ 1011.987189] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.994592] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062175, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.199322] env[61987]: INFO nova.compute.manager [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Unrescuing [ 1012.199613] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "refresh_cache-aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.199773] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquired lock "refresh_cache-aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.200016] env[61987]: DEBUG nova.network.neutron [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1012.228205] env[61987]: DEBUG nova.network.neutron [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Successfully updated port: 05cdb86d-67c4-4b34-9cd3-4e4b3ed04d58 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1012.242024] env[61987]: DEBUG nova.compute.manager [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1012.265617] env[61987]: DEBUG nova.compute.manager [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1012.266287] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1012.266796] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f21bd1-afcd-4b5e-9f3a-3da7a2be4819 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.275173] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1012.275466] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-427aeb56-ad6e-4d6b-a5d7-fc6cbc8f434a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.305667] env[61987]: DEBUG nova.compute.manager [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1012.363311] env[61987]: DEBUG oslo_concurrency.lockutils [req-70b561c6-d346-483e-bf66-7ccc57be6663 req-af3eda01-5dc2-4c06-8d51-e630d54865af service nova] Releasing lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.381491] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1012.381640] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1012.381786] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Deleting the datastore file [datastore1] 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1012.382332] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-81d52bda-800a-4ab5-a0f6-74902330f17e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.389320] env[61987]: DEBUG oslo_vmware.api [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 1012.389320] env[61987]: value = "task-1062177" [ 1012.389320] env[61987]: _type = "Task" [ 1012.389320] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.403360] env[61987]: DEBUG oslo_vmware.api [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062177, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.501013] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062175, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.608396] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-339be2ef-ce7b-42be-abab-dc6af32edb83 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.618254] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6138f808-cd85-4533-bab5-222aca68e609 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.661361] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-085fb3bf-604d-4c0c-bc19-7fce4cfc6f83 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.670414] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c680b6e-34bc-48cc-b2b5-fa381612866b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.685200] env[61987]: DEBUG nova.compute.provider_tree [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1012.731028] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquiring lock "refresh_cache-dab6d2c5-8c04-4977-a409-b384037eabce" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.731245] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquired lock "refresh_cache-dab6d2c5-8c04-4977-a409-b384037eabce" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.731428] env[61987]: DEBUG nova.network.neutron [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1012.763223] env[61987]: DEBUG oslo_concurrency.lockutils [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.824839] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.899958] env[61987]: DEBUG oslo_vmware.api [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062177, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.24577} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.900242] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1012.900450] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1012.900621] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1012.900801] env[61987]: INFO nova.compute.manager [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Took 0.63 seconds to destroy the instance on the hypervisor. [ 1012.901063] env[61987]: DEBUG oslo.service.loopingcall [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1012.901264] env[61987]: DEBUG nova.compute.manager [-] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1012.901363] env[61987]: DEBUG nova.network.neutron [-] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1012.998326] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062175, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.520628} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.998656] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 0da22c35-f11d-4498-bdb4-31bfd4c364e6/0da22c35-f11d-4498-bdb4-31bfd4c364e6.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1012.998971] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1012.999102] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bd1dcca5-5f8f-4824-87ff-a15666476dbf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.005914] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for the task: (returnval){ [ 1013.005914] env[61987]: value = "task-1062179" [ 1013.005914] env[61987]: _type = "Task" [ 1013.005914] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.013300] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062179, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.119835] env[61987]: DEBUG nova.network.neutron [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Updating instance_info_cache with network_info: [{"id": "70bf0473-599e-4d1e-be30-efec5a322e37", "address": "fa:16:3e:5e:ec:e5", "network": {"id": "4ed4fbdc-05b0-44f8-a3d5-b7b288abd5e0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2067313775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a234a62da0e498fbe613fbcaaec3201", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap70bf0473-59", "ovs_interfaceid": "70bf0473-599e-4d1e-be30-efec5a322e37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.150340] env[61987]: DEBUG nova.compute.manager [req-74452974-fe6b-45b7-8450-845d3c799828 req-dac4c626-e12c-4779-a646-96e66697833c service nova] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Received event network-vif-deleted-b3159788-bb87-42ad-b6e5-201ff2afe647 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1013.150548] env[61987]: INFO nova.compute.manager [req-74452974-fe6b-45b7-8450-845d3c799828 req-dac4c626-e12c-4779-a646-96e66697833c service nova] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Neutron deleted interface b3159788-bb87-42ad-b6e5-201ff2afe647; detaching it from the instance and deleting it from the info cache [ 1013.150727] env[61987]: DEBUG nova.network.neutron [req-74452974-fe6b-45b7-8450-845d3c799828 req-dac4c626-e12c-4779-a646-96e66697833c service nova] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.189563] env[61987]: DEBUG nova.scheduler.client.report [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1013.263088] env[61987]: DEBUG nova.network.neutron [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1013.307770] env[61987]: DEBUG nova.compute.manager [req-eefba160-39c6-41ad-b0d0-4cee1cb56b54 req-99bef7c6-956d-4c41-b92c-0808f154fee4 service nova] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Received event network-vif-plugged-05cdb86d-67c4-4b34-9cd3-4e4b3ed04d58 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1013.308120] env[61987]: DEBUG oslo_concurrency.lockutils [req-eefba160-39c6-41ad-b0d0-4cee1cb56b54 req-99bef7c6-956d-4c41-b92c-0808f154fee4 service nova] Acquiring lock "dab6d2c5-8c04-4977-a409-b384037eabce-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.308239] env[61987]: DEBUG oslo_concurrency.lockutils [req-eefba160-39c6-41ad-b0d0-4cee1cb56b54 req-99bef7c6-956d-4c41-b92c-0808f154fee4 service nova] Lock "dab6d2c5-8c04-4977-a409-b384037eabce-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.308663] env[61987]: DEBUG oslo_concurrency.lockutils [req-eefba160-39c6-41ad-b0d0-4cee1cb56b54 req-99bef7c6-956d-4c41-b92c-0808f154fee4 service nova] Lock "dab6d2c5-8c04-4977-a409-b384037eabce-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.308663] env[61987]: DEBUG nova.compute.manager [req-eefba160-39c6-41ad-b0d0-4cee1cb56b54 req-99bef7c6-956d-4c41-b92c-0808f154fee4 service nova] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] No waiting events found dispatching network-vif-plugged-05cdb86d-67c4-4b34-9cd3-4e4b3ed04d58 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1013.308849] env[61987]: WARNING nova.compute.manager [req-eefba160-39c6-41ad-b0d0-4cee1cb56b54 req-99bef7c6-956d-4c41-b92c-0808f154fee4 service nova] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Received unexpected event network-vif-plugged-05cdb86d-67c4-4b34-9cd3-4e4b3ed04d58 for instance with vm_state building and task_state spawning. [ 1013.309047] env[61987]: DEBUG nova.compute.manager [req-eefba160-39c6-41ad-b0d0-4cee1cb56b54 req-99bef7c6-956d-4c41-b92c-0808f154fee4 service nova] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Received event network-changed-05cdb86d-67c4-4b34-9cd3-4e4b3ed04d58 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1013.309212] env[61987]: DEBUG nova.compute.manager [req-eefba160-39c6-41ad-b0d0-4cee1cb56b54 req-99bef7c6-956d-4c41-b92c-0808f154fee4 service nova] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Refreshing instance network info cache due to event network-changed-05cdb86d-67c4-4b34-9cd3-4e4b3ed04d58. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1013.309378] env[61987]: DEBUG oslo_concurrency.lockutils [req-eefba160-39c6-41ad-b0d0-4cee1cb56b54 req-99bef7c6-956d-4c41-b92c-0808f154fee4 service nova] Acquiring lock "refresh_cache-dab6d2c5-8c04-4977-a409-b384037eabce" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.387096] env[61987]: DEBUG nova.network.neutron [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Updating instance_info_cache with network_info: [{"id": "05cdb86d-67c4-4b34-9cd3-4e4b3ed04d58", "address": "fa:16:3e:51:95:76", "network": {"id": "1157ea83-0804-47d3-96b7-7a5a16809f1a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-318208769-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "630439422a354de0b32ac5eefaa990b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69744f59-ecac-4b0b-831e-82a274d7acbb", "external-id": "nsx-vlan-transportzone-770", "segmentation_id": 770, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05cdb86d-67", "ovs_interfaceid": "05cdb86d-67c4-4b34-9cd3-4e4b3ed04d58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.517622] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062179, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.1133} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.517937] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1013.518968] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c04707cb-45ee-4fb8-a6ad-9cdff17abd65 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.544153] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 0da22c35-f11d-4498-bdb4-31bfd4c364e6/0da22c35-f11d-4498-bdb4-31bfd4c364e6.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1013.544620] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93fe2b0b-378e-4e61-9437-0ead45a74277 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.565918] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for the task: (returnval){ [ 1013.565918] env[61987]: value = "task-1062180" [ 1013.565918] env[61987]: _type = "Task" [ 1013.565918] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.575671] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062180, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.605418] env[61987]: DEBUG nova.network.neutron [-] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.622746] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Releasing lock "refresh_cache-aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.623602] env[61987]: DEBUG nova.objects.instance [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lazy-loading 'flavor' on Instance uuid aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.653675] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b44d93e0-199a-4073-94e2-2838e28b287c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.663400] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3a46642-d34e-4f8f-8bfd-ea9feaea32be {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.691098] env[61987]: DEBUG nova.compute.manager [req-74452974-fe6b-45b7-8450-845d3c799828 req-dac4c626-e12c-4779-a646-96e66697833c service nova] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Detach interface failed, port_id=b3159788-bb87-42ad-b6e5-201ff2afe647, reason: Instance 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 1013.696844] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.312s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.696844] env[61987]: DEBUG nova.compute.manager [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1013.698167] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.666s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.699089] env[61987]: DEBUG nova.objects.instance [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lazy-loading 'resources' on Instance uuid ae1396b3-c28b-4d1c-9a3a-3dc2170847ca {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.890071] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Releasing lock "refresh_cache-dab6d2c5-8c04-4977-a409-b384037eabce" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.890386] env[61987]: DEBUG nova.compute.manager [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Instance network_info: |[{"id": "05cdb86d-67c4-4b34-9cd3-4e4b3ed04d58", "address": "fa:16:3e:51:95:76", "network": {"id": "1157ea83-0804-47d3-96b7-7a5a16809f1a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-318208769-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "630439422a354de0b32ac5eefaa990b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69744f59-ecac-4b0b-831e-82a274d7acbb", "external-id": "nsx-vlan-transportzone-770", "segmentation_id": 770, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05cdb86d-67", "ovs_interfaceid": "05cdb86d-67c4-4b34-9cd3-4e4b3ed04d58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1013.890712] env[61987]: DEBUG oslo_concurrency.lockutils [req-eefba160-39c6-41ad-b0d0-4cee1cb56b54 req-99bef7c6-956d-4c41-b92c-0808f154fee4 service nova] Acquired lock "refresh_cache-dab6d2c5-8c04-4977-a409-b384037eabce" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.890899] env[61987]: DEBUG nova.network.neutron [req-eefba160-39c6-41ad-b0d0-4cee1cb56b54 req-99bef7c6-956d-4c41-b92c-0808f154fee4 service nova] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Refreshing network info cache for port 05cdb86d-67c4-4b34-9cd3-4e4b3ed04d58 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1013.892163] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:51:95:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69744f59-ecac-4b0b-831e-82a274d7acbb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '05cdb86d-67c4-4b34-9cd3-4e4b3ed04d58', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1013.899690] env[61987]: DEBUG oslo.service.loopingcall [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1013.902645] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1013.903149] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e9d09c10-258a-44ae-bc5d-269d15a89cf5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.923864] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1013.923864] env[61987]: value = "task-1062181" [ 1013.923864] env[61987]: _type = "Task" [ 1013.923864] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.932024] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062181, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.077801] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062180, 'name': ReconfigVM_Task, 'duration_secs': 0.284959} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.078162] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 0da22c35-f11d-4498-bdb4-31bfd4c364e6/0da22c35-f11d-4498-bdb4-31bfd4c364e6.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1014.078783] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0193cd03-c367-4c15-a646-6b4131b0b34b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.085093] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for the task: (returnval){ [ 1014.085093] env[61987]: value = "task-1062182" [ 1014.085093] env[61987]: _type = "Task" [ 1014.085093] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.093984] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062182, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.107911] env[61987]: INFO nova.compute.manager [-] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Took 1.21 seconds to deallocate network for instance. [ 1014.108935] env[61987]: DEBUG nova.network.neutron [req-eefba160-39c6-41ad-b0d0-4cee1cb56b54 req-99bef7c6-956d-4c41-b92c-0808f154fee4 service nova] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Updated VIF entry in instance network info cache for port 05cdb86d-67c4-4b34-9cd3-4e4b3ed04d58. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1014.110604] env[61987]: DEBUG nova.network.neutron [req-eefba160-39c6-41ad-b0d0-4cee1cb56b54 req-99bef7c6-956d-4c41-b92c-0808f154fee4 service nova] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Updating instance_info_cache with network_info: [{"id": "05cdb86d-67c4-4b34-9cd3-4e4b3ed04d58", "address": "fa:16:3e:51:95:76", "network": {"id": "1157ea83-0804-47d3-96b7-7a5a16809f1a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-318208769-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "630439422a354de0b32ac5eefaa990b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69744f59-ecac-4b0b-831e-82a274d7acbb", "external-id": "nsx-vlan-transportzone-770", "segmentation_id": 770, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05cdb86d-67", "ovs_interfaceid": "05cdb86d-67c4-4b34-9cd3-4e4b3ed04d58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.128900] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcaf5d97-9da1-4f9e-bcbd-f72ac1009b82 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.155306] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1014.155660] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-60d39de4-918d-42c9-9cbc-cf7fffa01d5f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.162934] env[61987]: DEBUG oslo_vmware.api [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1014.162934] env[61987]: value = "task-1062183" [ 1014.162934] env[61987]: _type = "Task" [ 1014.162934] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.171423] env[61987]: DEBUG oslo_vmware.api [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062183, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.201140] env[61987]: DEBUG nova.compute.utils [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1014.205418] env[61987]: DEBUG nova.compute.manager [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1014.205746] env[61987]: DEBUG nova.network.neutron [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1014.261504] env[61987]: DEBUG nova.policy [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2941054f63934469bf4daa6b5a912e14', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '764082416d314c3f92eb83f576aef222', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 1014.382050] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3fb0e7e-931a-4385-ad17-32a6dd446934 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.389736] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1543cab-a7e8-40df-9c12-8db1e63964fe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.418602] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0453acca-33dd-4335-aeb0-1239cc80ad6d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.429069] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b02979c-89bf-4874-a44d-69c9f172b482 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.439092] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062181, 'name': CreateVM_Task, 'duration_secs': 0.37943} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.446382] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1014.446882] env[61987]: DEBUG nova.compute.provider_tree [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.448455] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.448630] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.448956] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1014.449426] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-968e0b21-6698-4639-b368-777fc90f57cc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.453968] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1014.453968] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52810391-248d-7649-a844-202c8b3fa0b0" [ 1014.453968] env[61987]: _type = "Task" [ 1014.453968] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.461924] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52810391-248d-7649-a844-202c8b3fa0b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.513894] env[61987]: DEBUG nova.network.neutron [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Successfully created port: b5999bb0-480a-45bd-b07c-fbb91985421e {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1014.595625] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062182, 'name': Rename_Task, 'duration_secs': 0.19406} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.595860] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1014.596143] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a467c82-c34f-4110-be6b-f4ad35032bce {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.602966] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for the task: (returnval){ [ 1014.602966] env[61987]: value = "task-1062184" [ 1014.602966] env[61987]: _type = "Task" [ 1014.602966] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.611195] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062184, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.612812] env[61987]: DEBUG oslo_concurrency.lockutils [req-eefba160-39c6-41ad-b0d0-4cee1cb56b54 req-99bef7c6-956d-4c41-b92c-0808f154fee4 service nova] Releasing lock "refresh_cache-dab6d2c5-8c04-4977-a409-b384037eabce" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.615979] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.671982] env[61987]: DEBUG oslo_vmware.api [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062183, 'name': PowerOffVM_Task, 'duration_secs': 0.19274} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.672285] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1014.677774] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Reconfiguring VM instance instance-00000063 to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1014.678104] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c6dc582-8359-4a18-a409-fbb41ae56f30 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.696651] env[61987]: DEBUG oslo_vmware.api [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1014.696651] env[61987]: value = "task-1062185" [ 1014.696651] env[61987]: _type = "Task" [ 1014.696651] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.704290] env[61987]: DEBUG oslo_vmware.api [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062185, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.705902] env[61987]: DEBUG nova.compute.manager [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1014.843541] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "bf1123bf-80e6-4be3-804f-e13b906ee44d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.843793] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "bf1123bf-80e6-4be3-804f-e13b906ee44d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.950988] env[61987]: DEBUG nova.scheduler.client.report [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1014.964183] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52810391-248d-7649-a844-202c8b3fa0b0, 'name': SearchDatastore_Task, 'duration_secs': 0.034972} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.964486] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.964780] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1014.965061] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.965231] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.965418] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1014.965683] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0198b21d-e77e-48c5-84dd-f6f93b03e9e8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.974125] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1014.974311] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1014.975066] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da8b4be0-c506-4701-b69c-e4ef369456ab {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.980425] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1014.980425] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52af364e-6535-1554-2565-f195e5a63149" [ 1014.980425] env[61987]: _type = "Task" [ 1014.980425] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.987888] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52af364e-6535-1554-2565-f195e5a63149, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.112653] env[61987]: DEBUG oslo_vmware.api [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062184, 'name': PowerOnVM_Task, 'duration_secs': 0.459539} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.112653] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1015.112653] env[61987]: INFO nova.compute.manager [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Took 5.85 seconds to spawn the instance on the hypervisor. [ 1015.113127] env[61987]: DEBUG nova.compute.manager [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1015.113472] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59d623fe-6a25-4cfb-95cf-3fbb360268f9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.207056] env[61987]: DEBUG oslo_vmware.api [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062185, 'name': ReconfigVM_Task, 'duration_secs': 0.315461} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.207291] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Reconfigured VM instance instance-00000063 to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1015.207483] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1015.207730] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4e6a7cd1-a165-449c-a6d6-c4c3c137f616 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.218659] env[61987]: DEBUG oslo_vmware.api [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1015.218659] env[61987]: value = "task-1062186" [ 1015.218659] env[61987]: _type = "Task" [ 1015.218659] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.227698] env[61987]: DEBUG oslo_vmware.api [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062186, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.346248] env[61987]: DEBUG nova.compute.manager [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1015.458590] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.760s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.460900] env[61987]: DEBUG oslo_concurrency.lockutils [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.698s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.462363] env[61987]: INFO nova.compute.claims [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1015.478868] env[61987]: INFO nova.scheduler.client.report [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Deleted allocations for instance ae1396b3-c28b-4d1c-9a3a-3dc2170847ca [ 1015.492501] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52af364e-6535-1554-2565-f195e5a63149, 'name': SearchDatastore_Task, 'duration_secs': 0.012336} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.493352] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81affafe-ebeb-4054-b5a3-1e63e014dc31 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.498448] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1015.498448] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]523d7ca3-74a5-44db-542c-96319e57235d" [ 1015.498448] env[61987]: _type = "Task" [ 1015.498448] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.506628] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523d7ca3-74a5-44db-542c-96319e57235d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.630547] env[61987]: INFO nova.compute.manager [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Took 12.19 seconds to build instance. [ 1015.714809] env[61987]: DEBUG nova.compute.manager [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1015.728295] env[61987]: DEBUG oslo_vmware.api [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062186, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.744601] env[61987]: DEBUG nova.virt.hardware [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1015.744916] env[61987]: DEBUG nova.virt.hardware [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1015.745161] env[61987]: DEBUG nova.virt.hardware [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1015.745370] env[61987]: DEBUG nova.virt.hardware [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1015.745526] env[61987]: DEBUG nova.virt.hardware [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1015.745679] env[61987]: DEBUG nova.virt.hardware [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1015.745891] env[61987]: DEBUG nova.virt.hardware [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1015.746417] env[61987]: DEBUG nova.virt.hardware [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1015.746417] env[61987]: DEBUG nova.virt.hardware [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1015.746523] env[61987]: DEBUG nova.virt.hardware [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1015.746641] env[61987]: DEBUG nova.virt.hardware [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1015.747491] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc6a649e-5c26-49d2-93c5-d95960bbf139 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.755044] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f5c7b0f-4d88-40f9-b685-3a20c6b7f5a9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.873573] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.990031] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c20a5d83-ee44-440c-8b10-e7a3d9e0555e tempest-AttachVolumeTestJSON-1541715327 tempest-AttachVolumeTestJSON-1541715327-project-member] Lock "ae1396b3-c28b-4d1c-9a3a-3dc2170847ca" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.147s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.010818] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523d7ca3-74a5-44db-542c-96319e57235d, 'name': SearchDatastore_Task, 'duration_secs': 0.008801} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.011812] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.012049] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] dab6d2c5-8c04-4977-a409-b384037eabce/dab6d2c5-8c04-4977-a409-b384037eabce.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1016.014038] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7d280d14-b617-4923-b441-a8fd806aa4f2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.023778] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1016.023778] env[61987]: value = "task-1062187" [ 1016.023778] env[61987]: _type = "Task" [ 1016.023778] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.031060] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062187, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.058050] env[61987]: DEBUG nova.compute.manager [req-7804c72e-fdf8-4c27-846b-50ada4aeb8b5 req-1e7990f3-fbaf-4458-bdbb-32784e278075 service nova] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Received event network-vif-plugged-b5999bb0-480a-45bd-b07c-fbb91985421e {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1016.058745] env[61987]: DEBUG oslo_concurrency.lockutils [req-7804c72e-fdf8-4c27-846b-50ada4aeb8b5 req-1e7990f3-fbaf-4458-bdbb-32784e278075 service nova] Acquiring lock "f3773cda-41e8-4804-9319-c7e458112ddf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.059079] env[61987]: DEBUG oslo_concurrency.lockutils [req-7804c72e-fdf8-4c27-846b-50ada4aeb8b5 req-1e7990f3-fbaf-4458-bdbb-32784e278075 service nova] Lock "f3773cda-41e8-4804-9319-c7e458112ddf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.059181] env[61987]: DEBUG oslo_concurrency.lockutils [req-7804c72e-fdf8-4c27-846b-50ada4aeb8b5 req-1e7990f3-fbaf-4458-bdbb-32784e278075 service nova] Lock "f3773cda-41e8-4804-9319-c7e458112ddf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.059355] env[61987]: DEBUG nova.compute.manager [req-7804c72e-fdf8-4c27-846b-50ada4aeb8b5 req-1e7990f3-fbaf-4458-bdbb-32784e278075 service nova] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] No waiting events found dispatching network-vif-plugged-b5999bb0-480a-45bd-b07c-fbb91985421e {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1016.059549] env[61987]: WARNING nova.compute.manager [req-7804c72e-fdf8-4c27-846b-50ada4aeb8b5 req-1e7990f3-fbaf-4458-bdbb-32784e278075 service nova] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Received unexpected event network-vif-plugged-b5999bb0-480a-45bd-b07c-fbb91985421e for instance with vm_state building and task_state spawning. [ 1016.137401] env[61987]: DEBUG oslo_concurrency.lockutils [None req-495db570-7753-4a60-b0f8-62818a233b03 tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Lock "0da22c35-f11d-4498-bdb4-31bfd4c364e6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.708s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.167716] env[61987]: DEBUG nova.network.neutron [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Successfully updated port: b5999bb0-480a-45bd-b07c-fbb91985421e {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1016.229424] env[61987]: DEBUG oslo_vmware.api [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062186, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.536106] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062187, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.470688} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.539293] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] dab6d2c5-8c04-4977-a409-b384037eabce/dab6d2c5-8c04-4977-a409-b384037eabce.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1016.539730] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1016.540326] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ee9fb2eb-0176-44a9-be32-ae0b2e17cc3a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.547299] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1016.547299] env[61987]: value = "task-1062189" [ 1016.547299] env[61987]: _type = "Task" [ 1016.547299] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.558312] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062189, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.640372] env[61987]: INFO nova.compute.manager [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Rebuilding instance [ 1016.667332] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9eb6da4-d958-40ba-b3a3-d186a124f0f0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.673426] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "refresh_cache-f3773cda-41e8-4804-9319-c7e458112ddf" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.673567] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "refresh_cache-f3773cda-41e8-4804-9319-c7e458112ddf" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.673715] env[61987]: DEBUG nova.network.neutron [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1016.679588] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb40586-0891-4154-a9e0-e46c1318224a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.713014] env[61987]: DEBUG nova.compute.manager [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1016.713765] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-423d9f5b-ecfa-4f8f-8abe-087d3d4abce6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.716716] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2add3a3b-1853-4473-927f-09fedfcbd0f7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.726737] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b7b0e4e-c564-4b24-99b5-f3b57de84802 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.735571] env[61987]: DEBUG oslo_vmware.api [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062186, 'name': PowerOnVM_Task, 'duration_secs': 1.144635} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.736370] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1016.736617] env[61987]: DEBUG nova.compute.manager [None req-d61b86cf-9663-442a-9e53-dba1e7bcabb4 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1016.737501] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ecbe2b9-28db-4fdc-bcdb-c4f6979fb399 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.747348] env[61987]: DEBUG nova.compute.provider_tree [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.061317] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062189, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.106514} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.061589] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1017.066320] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12113da7-1068-4c63-8299-4943d4ee8c1c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.090021] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] dab6d2c5-8c04-4977-a409-b384037eabce/dab6d2c5-8c04-4977-a409-b384037eabce.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1017.091921] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1f9688f0-ba3d-4f1d-aa5e-2b5fba93c8f7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.111764] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1017.111764] env[61987]: value = "task-1062190" [ 1017.111764] env[61987]: _type = "Task" [ 1017.111764] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.119673] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062190, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.206664] env[61987]: DEBUG nova.network.neutron [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1017.249832] env[61987]: DEBUG nova.scheduler.client.report [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1017.362446] env[61987]: DEBUG nova.network.neutron [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Updating instance_info_cache with network_info: [{"id": "b5999bb0-480a-45bd-b07c-fbb91985421e", "address": "fa:16:3e:00:34:7e", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5999bb0-48", "ovs_interfaceid": "b5999bb0-480a-45bd-b07c-fbb91985421e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.624778] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062190, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.740640] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1017.740945] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca0dea9f-2643-4ece-b402-c4e68fc0d03d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.748641] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for the task: (returnval){ [ 1017.748641] env[61987]: value = "task-1062191" [ 1017.748641] env[61987]: _type = "Task" [ 1017.748641] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.758421] env[61987]: DEBUG oslo_concurrency.lockutils [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.297s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.759344] env[61987]: DEBUG nova.compute.manager [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1017.761493] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062191, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.761984] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.937s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.763564] env[61987]: INFO nova.compute.claims [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1017.865232] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "refresh_cache-f3773cda-41e8-4804-9319-c7e458112ddf" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.866072] env[61987]: DEBUG nova.compute.manager [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Instance network_info: |[{"id": "b5999bb0-480a-45bd-b07c-fbb91985421e", "address": "fa:16:3e:00:34:7e", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5999bb0-48", "ovs_interfaceid": "b5999bb0-480a-45bd-b07c-fbb91985421e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1017.866072] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:00:34:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2a111da1-447a-4722-9662-b0d6a2886a65', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b5999bb0-480a-45bd-b07c-fbb91985421e', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1017.875539] env[61987]: DEBUG oslo.service.loopingcall [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1017.876196] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1017.876491] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e13f31e7-487f-424c-974e-6ea927674dbc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.896444] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1017.896444] env[61987]: value = "task-1062192" [ 1017.896444] env[61987]: _type = "Task" [ 1017.896444] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.904196] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062192, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.122511] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062190, 'name': ReconfigVM_Task, 'duration_secs': 0.953834} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.122927] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Reconfigured VM instance instance-00000069 to attach disk [datastore1] dab6d2c5-8c04-4977-a409-b384037eabce/dab6d2c5-8c04-4977-a409-b384037eabce.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1018.123538] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8c721dc6-680d-4eda-b2ab-84f69e50283d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.129833] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1018.129833] env[61987]: value = "task-1062193" [ 1018.129833] env[61987]: _type = "Task" [ 1018.129833] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.140366] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062193, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.204560] env[61987]: DEBUG nova.compute.manager [req-0c5a0cd2-64d1-4782-b403-9efc023659d2 req-5f705f8a-00d7-4008-ac24-bc07f2eeca0a service nova] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Received event network-changed-b5999bb0-480a-45bd-b07c-fbb91985421e {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1018.204859] env[61987]: DEBUG nova.compute.manager [req-0c5a0cd2-64d1-4782-b403-9efc023659d2 req-5f705f8a-00d7-4008-ac24-bc07f2eeca0a service nova] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Refreshing instance network info cache due to event network-changed-b5999bb0-480a-45bd-b07c-fbb91985421e. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1018.205182] env[61987]: DEBUG oslo_concurrency.lockutils [req-0c5a0cd2-64d1-4782-b403-9efc023659d2 req-5f705f8a-00d7-4008-ac24-bc07f2eeca0a service nova] Acquiring lock "refresh_cache-f3773cda-41e8-4804-9319-c7e458112ddf" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.205357] env[61987]: DEBUG oslo_concurrency.lockutils [req-0c5a0cd2-64d1-4782-b403-9efc023659d2 req-5f705f8a-00d7-4008-ac24-bc07f2eeca0a service nova] Acquired lock "refresh_cache-f3773cda-41e8-4804-9319-c7e458112ddf" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.205535] env[61987]: DEBUG nova.network.neutron [req-0c5a0cd2-64d1-4782-b403-9efc023659d2 req-5f705f8a-00d7-4008-ac24-bc07f2eeca0a service nova] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Refreshing network info cache for port b5999bb0-480a-45bd-b07c-fbb91985421e {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1018.258648] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062191, 'name': PowerOffVM_Task, 'duration_secs': 0.151342} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.258956] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1018.259225] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1018.260012] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09eae92c-ba56-4d70-8cac-312c12f81d5f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.264265] env[61987]: DEBUG nova.compute.utils [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1018.265713] env[61987]: DEBUG nova.compute.manager [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1018.265895] env[61987]: DEBUG nova.network.neutron [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1018.274331] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1018.274738] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b956bf89-7071-4589-9777-63af048d61b1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.310970] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1018.311481] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1018.311877] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Deleting the datastore file [datastore1] 0da22c35-f11d-4498-bdb4-31bfd4c364e6 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1018.312683] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4d5beacd-1c3c-48af-be12-41902f0a9772 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.320386] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for the task: (returnval){ [ 1018.320386] env[61987]: value = "task-1062196" [ 1018.320386] env[61987]: _type = "Task" [ 1018.320386] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.326799] env[61987]: DEBUG nova.policy [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c5cf7a7673cc4f5aadca7960c355c394', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ea5c65ac3b143178b13731663dbd8ee', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 1018.334834] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062196, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.408598] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062192, 'name': CreateVM_Task, 'duration_secs': 0.328345} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.408799] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1018.409922] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.410183] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.410544] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1018.410809] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8435db2-31e2-463b-84b3-4dcfb9ee4ecd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.415603] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1018.415603] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52d8ffd1-7b03-8d8f-5500-bcc5b3825123" [ 1018.415603] env[61987]: _type = "Task" [ 1018.415603] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.424560] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52d8ffd1-7b03-8d8f-5500-bcc5b3825123, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.576840] env[61987]: DEBUG nova.network.neutron [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Successfully created port: 2a9d7219-a935-4e81-9278-7abce87993fe {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1018.642151] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062193, 'name': Rename_Task, 'duration_secs': 0.194302} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.642506] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1018.642988] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aecb08e2-08da-43f9-a143-a987a10b23df {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.649485] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1018.649485] env[61987]: value = "task-1062197" [ 1018.649485] env[61987]: _type = "Task" [ 1018.649485] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.659053] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062197, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.769507] env[61987]: DEBUG nova.compute.manager [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1018.833561] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062196, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.222327} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.837668] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1018.837898] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1018.838102] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1018.926342] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52d8ffd1-7b03-8d8f-5500-bcc5b3825123, 'name': SearchDatastore_Task, 'duration_secs': 0.032212} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.929576] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.929847] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1018.930098] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.930258] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.930445] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1018.930894] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a4837ea-9852-4d03-b09b-ff11032eb53b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.939334] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1018.939719] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1018.943250] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-074da607-b96f-47fa-86fc-af20ad7462c5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.950392] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1018.950392] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52db61cf-58c7-86d6-d393-e535c5f3b1fe" [ 1018.950392] env[61987]: _type = "Task" [ 1018.950392] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.956859] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52db61cf-58c7-86d6-d393-e535c5f3b1fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.963544] env[61987]: DEBUG nova.network.neutron [req-0c5a0cd2-64d1-4782-b403-9efc023659d2 req-5f705f8a-00d7-4008-ac24-bc07f2eeca0a service nova] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Updated VIF entry in instance network info cache for port b5999bb0-480a-45bd-b07c-fbb91985421e. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1018.964039] env[61987]: DEBUG nova.network.neutron [req-0c5a0cd2-64d1-4782-b403-9efc023659d2 req-5f705f8a-00d7-4008-ac24-bc07f2eeca0a service nova] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Updating instance_info_cache with network_info: [{"id": "b5999bb0-480a-45bd-b07c-fbb91985421e", "address": "fa:16:3e:00:34:7e", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5999bb0-48", "ovs_interfaceid": "b5999bb0-480a-45bd-b07c-fbb91985421e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.996374] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20faeb75-4644-4eaa-8571-5463d010fd2d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.004272] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-978a10db-fbce-4414-be5f-5a682210295e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.036472] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-489e12c7-e7fb-488e-aef7-5ff592a8f6eb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.044602] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-385a679c-533c-4cf9-b02a-d26a63c54012 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.058876] env[61987]: DEBUG nova.compute.provider_tree [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.159453] env[61987]: DEBUG oslo_vmware.api [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062197, 'name': PowerOnVM_Task, 'duration_secs': 0.482695} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.159732] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1019.159940] env[61987]: INFO nova.compute.manager [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Took 7.56 seconds to spawn the instance on the hypervisor. [ 1019.160148] env[61987]: DEBUG nova.compute.manager [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1019.160943] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9ed602b-1069-409e-8066-a0bcbdc90867 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.459527] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52db61cf-58c7-86d6-d393-e535c5f3b1fe, 'name': SearchDatastore_Task, 'duration_secs': 0.010567} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.460561] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c172a9a-b002-47f8-b6b2-60dc71c42f3a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.465817] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1019.465817] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52c83f17-ab14-8585-8c2b-e4be6ccf4b82" [ 1019.465817] env[61987]: _type = "Task" [ 1019.465817] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.466455] env[61987]: DEBUG oslo_concurrency.lockutils [req-0c5a0cd2-64d1-4782-b403-9efc023659d2 req-5f705f8a-00d7-4008-ac24-bc07f2eeca0a service nova] Releasing lock "refresh_cache-f3773cda-41e8-4804-9319-c7e458112ddf" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.479139] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c83f17-ab14-8585-8c2b-e4be6ccf4b82, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.561853] env[61987]: DEBUG nova.scheduler.client.report [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1019.679033] env[61987]: INFO nova.compute.manager [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Took 13.86 seconds to build instance. [ 1019.782508] env[61987]: DEBUG nova.compute.manager [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1019.804329] env[61987]: DEBUG nova.virt.hardware [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1019.804628] env[61987]: DEBUG nova.virt.hardware [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1019.804822] env[61987]: DEBUG nova.virt.hardware [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1019.805114] env[61987]: DEBUG nova.virt.hardware [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1019.805322] env[61987]: DEBUG nova.virt.hardware [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1019.805515] env[61987]: DEBUG nova.virt.hardware [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1019.805752] env[61987]: DEBUG nova.virt.hardware [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1019.806372] env[61987]: DEBUG nova.virt.hardware [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1019.806372] env[61987]: DEBUG nova.virt.hardware [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1019.806474] env[61987]: DEBUG nova.virt.hardware [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1019.806725] env[61987]: DEBUG nova.virt.hardware [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1019.807666] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5574cd71-00a7-46f5-b332-c7a34da00dde {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.815700] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22c497f2-0bd9-4a48-9292-20c6369568d6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.868755] env[61987]: DEBUG nova.virt.hardware [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1019.869102] env[61987]: DEBUG nova.virt.hardware [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1019.869326] env[61987]: DEBUG nova.virt.hardware [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1019.869544] env[61987]: DEBUG nova.virt.hardware [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1019.869703] env[61987]: DEBUG nova.virt.hardware [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1019.869857] env[61987]: DEBUG nova.virt.hardware [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1019.870080] env[61987]: DEBUG nova.virt.hardware [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1019.870255] env[61987]: DEBUG nova.virt.hardware [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1019.870430] env[61987]: DEBUG nova.virt.hardware [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1019.870615] env[61987]: DEBUG nova.virt.hardware [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1019.870820] env[61987]: DEBUG nova.virt.hardware [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1019.871669] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d754b7d-7439-4519-97da-fd29bdb43d1c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.879058] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-865e763a-f84f-4f9e-83d4-04ad199453c7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.893797] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Instance VIF info [] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1019.899508] env[61987]: DEBUG oslo.service.loopingcall [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1019.899685] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1019.899893] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-018f399b-8d58-460e-b3e6-2e8abe4442b6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.919467] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1019.919467] env[61987]: value = "task-1062198" [ 1019.919467] env[61987]: _type = "Task" [ 1019.919467] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.927207] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062198, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.976715] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c83f17-ab14-8585-8c2b-e4be6ccf4b82, 'name': SearchDatastore_Task, 'duration_secs': 0.030481} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.976715] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.976882] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] f3773cda-41e8-4804-9319-c7e458112ddf/f3773cda-41e8-4804-9319-c7e458112ddf.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1019.977167] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e6129e12-913d-4fa8-bca6-48b6379fb703 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.983256] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1019.983256] env[61987]: value = "task-1062199" [ 1019.983256] env[61987]: _type = "Task" [ 1019.983256] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.991100] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062199, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.067458] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.305s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.068200] env[61987]: DEBUG nova.compute.manager [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1020.072082] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.456s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.072388] env[61987]: DEBUG nova.objects.instance [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lazy-loading 'resources' on Instance uuid 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1020.145309] env[61987]: DEBUG nova.network.neutron [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Successfully updated port: 2a9d7219-a935-4e81-9278-7abce87993fe {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1020.181643] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3d6c8668-1f21-4cc9-9716-579bd1e7d508 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lock "dab6d2c5-8c04-4977-a409-b384037eabce" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.366s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.239623] env[61987]: DEBUG nova.compute.manager [req-52ef45d3-5db5-4c53-a79f-45151738add3 req-377c891d-0cd4-419e-a35c-4f314138c8c9 service nova] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Received event network-vif-plugged-2a9d7219-a935-4e81-9278-7abce87993fe {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1020.239829] env[61987]: DEBUG oslo_concurrency.lockutils [req-52ef45d3-5db5-4c53-a79f-45151738add3 req-377c891d-0cd4-419e-a35c-4f314138c8c9 service nova] Acquiring lock "b172195a-79d2-4f24-b9df-320a07de965e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.240251] env[61987]: DEBUG oslo_concurrency.lockutils [req-52ef45d3-5db5-4c53-a79f-45151738add3 req-377c891d-0cd4-419e-a35c-4f314138c8c9 service nova] Lock "b172195a-79d2-4f24-b9df-320a07de965e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.240478] env[61987]: DEBUG oslo_concurrency.lockutils [req-52ef45d3-5db5-4c53-a79f-45151738add3 req-377c891d-0cd4-419e-a35c-4f314138c8c9 service nova] Lock "b172195a-79d2-4f24-b9df-320a07de965e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.240666] env[61987]: DEBUG nova.compute.manager [req-52ef45d3-5db5-4c53-a79f-45151738add3 req-377c891d-0cd4-419e-a35c-4f314138c8c9 service nova] [instance: b172195a-79d2-4f24-b9df-320a07de965e] No waiting events found dispatching network-vif-plugged-2a9d7219-a935-4e81-9278-7abce87993fe {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1020.240896] env[61987]: WARNING nova.compute.manager [req-52ef45d3-5db5-4c53-a79f-45151738add3 req-377c891d-0cd4-419e-a35c-4f314138c8c9 service nova] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Received unexpected event network-vif-plugged-2a9d7219-a935-4e81-9278-7abce87993fe for instance with vm_state building and task_state spawning. [ 1020.241154] env[61987]: DEBUG nova.compute.manager [req-52ef45d3-5db5-4c53-a79f-45151738add3 req-377c891d-0cd4-419e-a35c-4f314138c8c9 service nova] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Received event network-changed-2a9d7219-a935-4e81-9278-7abce87993fe {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1020.241343] env[61987]: DEBUG nova.compute.manager [req-52ef45d3-5db5-4c53-a79f-45151738add3 req-377c891d-0cd4-419e-a35c-4f314138c8c9 service nova] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Refreshing instance network info cache due to event network-changed-2a9d7219-a935-4e81-9278-7abce87993fe. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1020.241594] env[61987]: DEBUG oslo_concurrency.lockutils [req-52ef45d3-5db5-4c53-a79f-45151738add3 req-377c891d-0cd4-419e-a35c-4f314138c8c9 service nova] Acquiring lock "refresh_cache-b172195a-79d2-4f24-b9df-320a07de965e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.241747] env[61987]: DEBUG oslo_concurrency.lockutils [req-52ef45d3-5db5-4c53-a79f-45151738add3 req-377c891d-0cd4-419e-a35c-4f314138c8c9 service nova] Acquired lock "refresh_cache-b172195a-79d2-4f24-b9df-320a07de965e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.241936] env[61987]: DEBUG nova.network.neutron [req-52ef45d3-5db5-4c53-a79f-45151738add3 req-377c891d-0cd4-419e-a35c-4f314138c8c9 service nova] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Refreshing network info cache for port 2a9d7219-a935-4e81-9278-7abce87993fe {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1020.430435] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062198, 'name': CreateVM_Task, 'duration_secs': 0.330204} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.430750] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1020.431173] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.431371] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.431702] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1020.431964] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b463a695-b206-4ff7-ac6a-b33848684248 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.437029] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for the task: (returnval){ [ 1020.437029] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52b14ac1-68da-0282-f0a0-5efbe005c693" [ 1020.437029] env[61987]: _type = "Task" [ 1020.437029] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.444575] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b14ac1-68da-0282-f0a0-5efbe005c693, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.492638] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062199, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450851} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.492938] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] f3773cda-41e8-4804-9319-c7e458112ddf/f3773cda-41e8-4804-9319-c7e458112ddf.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1020.493176] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1020.493500] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cdd30173-8735-4b52-b15e-471f0194e497 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.500268] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1020.500268] env[61987]: value = "task-1062200" [ 1020.500268] env[61987]: _type = "Task" [ 1020.500268] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.508243] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062200, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.575634] env[61987]: DEBUG nova.compute.utils [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1020.579431] env[61987]: DEBUG nova.compute.manager [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1020.580025] env[61987]: DEBUG nova.network.neutron [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1020.649023] env[61987]: DEBUG oslo_concurrency.lockutils [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "refresh_cache-b172195a-79d2-4f24-b9df-320a07de965e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.656217] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "7002962c-76ed-4a8d-be2a-b2b118847fea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.656427] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "7002962c-76ed-4a8d-be2a-b2b118847fea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.662526] env[61987]: DEBUG nova.policy [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '697f9b1e3b3840369549a4f1ae3173ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd2afdc9bddba4ccab2b8fdb1f31da2b6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 1020.766866] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf17a799-4944-41eb-a0c5-974ecc6df7c6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.774421] env[61987]: DEBUG nova.network.neutron [req-52ef45d3-5db5-4c53-a79f-45151738add3 req-377c891d-0cd4-419e-a35c-4f314138c8c9 service nova] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1020.776864] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39935444-881f-4ad8-9bc6-9f42dc75eae3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.807154] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4b4bf5-3d80-434d-af6f-a64694c1253d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.814382] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684b1c2a-b7d6-43d4-a01d-31b11940ed81 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.830706] env[61987]: DEBUG nova.compute.provider_tree [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1020.854440] env[61987]: DEBUG nova.network.neutron [req-52ef45d3-5db5-4c53-a79f-45151738add3 req-377c891d-0cd4-419e-a35c-4f314138c8c9 service nova] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.946623] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b14ac1-68da-0282-f0a0-5efbe005c693, 'name': SearchDatastore_Task, 'duration_secs': 0.009931} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.946944] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.947197] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1020.947439] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.947594] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.947794] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1020.948070] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e11155f9-666f-407d-99ad-624d08ffe16c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.954300] env[61987]: DEBUG nova.network.neutron [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Successfully created port: b739a840-515d-46e6-b889-5e9f8ec9ad71 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1020.957809] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1020.958079] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1020.959709] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a219b37b-e1c1-4123-8328-ed88f3e184ea {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.964049] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for the task: (returnval){ [ 1020.964049] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52ba55e0-2a73-fbe6-5c38-c42a05844e20" [ 1020.964049] env[61987]: _type = "Task" [ 1020.964049] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.972450] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52ba55e0-2a73-fbe6-5c38-c42a05844e20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.009725] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062200, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067688} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.010014] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1021.010799] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01627916-aac9-4031-aaff-d9942a32d0f3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.032917] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] f3773cda-41e8-4804-9319-c7e458112ddf/f3773cda-41e8-4804-9319-c7e458112ddf.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1021.033243] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3856b01a-9ba7-4c9b-a6d4-79ffa6342548 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.053799] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1021.053799] env[61987]: value = "task-1062201" [ 1021.053799] env[61987]: _type = "Task" [ 1021.053799] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.062132] env[61987]: INFO nova.compute.manager [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Rescuing [ 1021.062381] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquiring lock "refresh_cache-dab6d2c5-8c04-4977-a409-b384037eabce" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.062567] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquired lock "refresh_cache-dab6d2c5-8c04-4977-a409-b384037eabce" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.062740] env[61987]: DEBUG nova.network.neutron [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1021.063957] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062201, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.080251] env[61987]: DEBUG nova.compute.manager [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1021.162869] env[61987]: DEBUG nova.compute.manager [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1021.333682] env[61987]: DEBUG nova.scheduler.client.report [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1021.356636] env[61987]: DEBUG oslo_concurrency.lockutils [req-52ef45d3-5db5-4c53-a79f-45151738add3 req-377c891d-0cd4-419e-a35c-4f314138c8c9 service nova] Releasing lock "refresh_cache-b172195a-79d2-4f24-b9df-320a07de965e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.357300] env[61987]: DEBUG oslo_concurrency.lockutils [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired lock "refresh_cache-b172195a-79d2-4f24-b9df-320a07de965e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.357472] env[61987]: DEBUG nova.network.neutron [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1021.475792] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52ba55e0-2a73-fbe6-5c38-c42a05844e20, 'name': SearchDatastore_Task, 'duration_secs': 0.053538} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.476691] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c483bc2d-dc5b-4fc0-831a-6eb0669e3056 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.481867] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for the task: (returnval){ [ 1021.481867] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]526c70e9-ca09-5c25-c924-8f86b27324b4" [ 1021.481867] env[61987]: _type = "Task" [ 1021.481867] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.489661] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]526c70e9-ca09-5c25-c924-8f86b27324b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.563326] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062201, 'name': ReconfigVM_Task, 'duration_secs': 0.332407} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.563652] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Reconfigured VM instance instance-0000006a to attach disk [datastore1] f3773cda-41e8-4804-9319-c7e458112ddf/f3773cda-41e8-4804-9319-c7e458112ddf.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1021.564348] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc04f2e6-e0df-43ff-9b51-27477ec2f80f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.572283] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1021.572283] env[61987]: value = "task-1062203" [ 1021.572283] env[61987]: _type = "Task" [ 1021.572283] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.582274] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062203, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.690332] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.818898] env[61987]: DEBUG nova.network.neutron [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Updating instance_info_cache with network_info: [{"id": "05cdb86d-67c4-4b34-9cd3-4e4b3ed04d58", "address": "fa:16:3e:51:95:76", "network": {"id": "1157ea83-0804-47d3-96b7-7a5a16809f1a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-318208769-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "630439422a354de0b32ac5eefaa990b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69744f59-ecac-4b0b-831e-82a274d7acbb", "external-id": "nsx-vlan-transportzone-770", "segmentation_id": 770, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05cdb86d-67", "ovs_interfaceid": "05cdb86d-67c4-4b34-9cd3-4e4b3ed04d58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.838395] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.766s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.840649] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.967s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.842358] env[61987]: INFO nova.compute.claims [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1021.857298] env[61987]: INFO nova.scheduler.client.report [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Deleted allocations for instance 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4 [ 1021.895502] env[61987]: DEBUG nova.network.neutron [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1021.992070] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]526c70e9-ca09-5c25-c924-8f86b27324b4, 'name': SearchDatastore_Task, 'duration_secs': 0.020294} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.992355] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.992621] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 0da22c35-f11d-4498-bdb4-31bfd4c364e6/0da22c35-f11d-4498-bdb4-31bfd4c364e6.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1021.992909] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ed54b8d8-cc83-4aa9-9fcd-a2ec0e1ac82e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.999620] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for the task: (returnval){ [ 1021.999620] env[61987]: value = "task-1062204" [ 1021.999620] env[61987]: _type = "Task" [ 1021.999620] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.007244] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062204, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.033265] env[61987]: DEBUG nova.network.neutron [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Updating instance_info_cache with network_info: [{"id": "2a9d7219-a935-4e81-9278-7abce87993fe", "address": "fa:16:3e:20:28:cc", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a9d7219-a9", "ovs_interfaceid": "2a9d7219-a935-4e81-9278-7abce87993fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.081993] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062203, 'name': Rename_Task, 'duration_secs': 0.163242} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.082608] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1022.082888] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-83afc702-a2db-4cd1-ad93-8a9fdef73b63 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.089390] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1022.089390] env[61987]: value = "task-1062205" [ 1022.089390] env[61987]: _type = "Task" [ 1022.089390] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.093386] env[61987]: DEBUG nova.compute.manager [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1022.100130] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062205, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.120307] env[61987]: DEBUG nova.virt.hardware [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1022.120595] env[61987]: DEBUG nova.virt.hardware [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1022.120787] env[61987]: DEBUG nova.virt.hardware [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1022.121079] env[61987]: DEBUG nova.virt.hardware [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1022.121326] env[61987]: DEBUG nova.virt.hardware [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1022.121523] env[61987]: DEBUG nova.virt.hardware [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1022.121745] env[61987]: DEBUG nova.virt.hardware [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1022.121916] env[61987]: DEBUG nova.virt.hardware [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1022.122106] env[61987]: DEBUG nova.virt.hardware [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1022.122283] env[61987]: DEBUG nova.virt.hardware [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1022.122467] env[61987]: DEBUG nova.virt.hardware [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1022.123351] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c73cc8a-f62f-410a-b0e6-22e2bf2986d3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.131254] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e90ef7d-8cee-436c-99c0-4f1943651cf9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.321633] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Releasing lock "refresh_cache-dab6d2c5-8c04-4977-a409-b384037eabce" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.367939] env[61987]: DEBUG nova.compute.manager [req-4b156d8d-6f63-43a4-a309-79150762cd8f req-c025b77e-f37e-4b72-8312-0aa53943818b service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Received event network-vif-plugged-b739a840-515d-46e6-b889-5e9f8ec9ad71 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1022.368180] env[61987]: DEBUG oslo_concurrency.lockutils [req-4b156d8d-6f63-43a4-a309-79150762cd8f req-c025b77e-f37e-4b72-8312-0aa53943818b service nova] Acquiring lock "0cba5e51-7e0a-409b-a7b2-612ab8cf5689-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.368398] env[61987]: DEBUG oslo_concurrency.lockutils [req-4b156d8d-6f63-43a4-a309-79150762cd8f req-c025b77e-f37e-4b72-8312-0aa53943818b service nova] Lock "0cba5e51-7e0a-409b-a7b2-612ab8cf5689-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.368717] env[61987]: DEBUG oslo_concurrency.lockutils [req-4b156d8d-6f63-43a4-a309-79150762cd8f req-c025b77e-f37e-4b72-8312-0aa53943818b service nova] Lock "0cba5e51-7e0a-409b-a7b2-612ab8cf5689-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.368780] env[61987]: DEBUG nova.compute.manager [req-4b156d8d-6f63-43a4-a309-79150762cd8f req-c025b77e-f37e-4b72-8312-0aa53943818b service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] No waiting events found dispatching network-vif-plugged-b739a840-515d-46e6-b889-5e9f8ec9ad71 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1022.368959] env[61987]: WARNING nova.compute.manager [req-4b156d8d-6f63-43a4-a309-79150762cd8f req-c025b77e-f37e-4b72-8312-0aa53943818b service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Received unexpected event network-vif-plugged-b739a840-515d-46e6-b889-5e9f8ec9ad71 for instance with vm_state building and task_state spawning. [ 1022.369383] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5e99fd12-847d-4a00-b617-f1bcf6dd2c41 tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "54d9bbcb-d01c-4e3b-8feb-4f8942d475d4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.611s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.446345] env[61987]: DEBUG nova.network.neutron [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Successfully updated port: b739a840-515d-46e6-b889-5e9f8ec9ad71 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1022.510034] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062204, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.536541] env[61987]: DEBUG oslo_concurrency.lockutils [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Releasing lock "refresh_cache-b172195a-79d2-4f24-b9df-320a07de965e" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.536963] env[61987]: DEBUG nova.compute.manager [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Instance network_info: |[{"id": "2a9d7219-a935-4e81-9278-7abce87993fe", "address": "fa:16:3e:20:28:cc", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2a9d7219-a9", "ovs_interfaceid": "2a9d7219-a935-4e81-9278-7abce87993fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1022.537546] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:20:28:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'be8bd197-4b2b-46e7-88ea-2554b0438584', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2a9d7219-a935-4e81-9278-7abce87993fe', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1022.545998] env[61987]: DEBUG oslo.service.loopingcall [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.546211] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1022.546459] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2233a64e-b920-4c04-81b7-ca9be4869ad7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.576135] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1022.576135] env[61987]: value = "task-1062206" [ 1022.576135] env[61987]: _type = "Task" [ 1022.576135] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.587279] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062206, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.598426] env[61987]: DEBUG oslo_vmware.api [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062205, 'name': PowerOnVM_Task, 'duration_secs': 0.503708} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.598727] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1022.598945] env[61987]: INFO nova.compute.manager [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Took 6.88 seconds to spawn the instance on the hypervisor. [ 1022.599213] env[61987]: DEBUG nova.compute.manager [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1022.600126] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2153f59-5a8f-415c-a8ad-ab861754a244 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.950641] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.950811] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.950957] env[61987]: DEBUG nova.network.neutron [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1023.010062] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062204, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.806171} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.010797] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 0da22c35-f11d-4498-bdb4-31bfd4c364e6/0da22c35-f11d-4498-bdb4-31bfd4c364e6.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1023.010797] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1023.010797] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-02f29f59-09f5-479f-9246-04ff00a584fd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.017112] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for the task: (returnval){ [ 1023.017112] env[61987]: value = "task-1062207" [ 1023.017112] env[61987]: _type = "Task" [ 1023.017112] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.025202] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062207, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.049559] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c41299-3d06-4bf8-8104-ceab818ed8b7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.057960] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5628f688-03c9-4128-b751-02f617795ca7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.091887] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce635c29-3156-4d6e-9927-10c0c0981ce2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.099695] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062206, 'name': CreateVM_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.104179] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcf9293c-d508-4507-859a-b779166589cc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.118102] env[61987]: INFO nova.compute.manager [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Took 12.63 seconds to build instance. [ 1023.127714] env[61987]: DEBUG nova.compute.provider_tree [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1023.484700] env[61987]: DEBUG nova.network.neutron [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1023.529514] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062207, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.255845} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.529799] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1023.530830] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf36889b-f98b-4a63-95aa-6963e4acd9b9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.551658] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 0da22c35-f11d-4498-bdb4-31bfd4c364e6/0da22c35-f11d-4498-bdb4-31bfd4c364e6.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1023.554543] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-113e751e-48d8-43f6-abae-f1e552c61fbe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.575187] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for the task: (returnval){ [ 1023.575187] env[61987]: value = "task-1062208" [ 1023.575187] env[61987]: _type = "Task" [ 1023.575187] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.583363] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062208, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.596802] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062206, 'name': CreateVM_Task, 'duration_secs': 0.956952} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.596986] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1023.597732] env[61987]: DEBUG oslo_concurrency.lockutils [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.597929] env[61987]: DEBUG oslo_concurrency.lockutils [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.598280] env[61987]: DEBUG oslo_concurrency.lockutils [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1023.598543] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77d229c0-7754-40f4-af83-115d2842c248 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.603392] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1023.603392] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52743964-164d-7dbc-e3ce-d3f280a07871" [ 1023.603392] env[61987]: _type = "Task" [ 1023.603392] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.613796] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "84a4a69b-04db-4546-ba89-9039e382a0c4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.614693] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "84a4a69b-04db-4546-ba89-9039e382a0c4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.614693] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "84a4a69b-04db-4546-ba89-9039e382a0c4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.615028] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "84a4a69b-04db-4546-ba89-9039e382a0c4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.615268] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "84a4a69b-04db-4546-ba89-9039e382a0c4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.617187] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52743964-164d-7dbc-e3ce-d3f280a07871, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.617550] env[61987]: INFO nova.compute.manager [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Terminating instance [ 1023.619531] env[61987]: DEBUG oslo_concurrency.lockutils [None req-3874d25a-57fe-4dce-bcd3-97b4ee13ff7b tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "f3773cda-41e8-4804-9319-c7e458112ddf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.147s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.630756] env[61987]: DEBUG nova.scheduler.client.report [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1023.860227] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1023.860948] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c6847cb1-bda5-4f67-8f2c-b6630222837a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.871386] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1023.871386] env[61987]: value = "task-1062209" [ 1023.871386] env[61987]: _type = "Task" [ 1023.871386] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.883847] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062209, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.898746] env[61987]: DEBUG nova.network.neutron [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Updating instance_info_cache with network_info: [{"id": "b739a840-515d-46e6-b889-5e9f8ec9ad71", "address": "fa:16:3e:cb:9f:9b", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb739a840-51", "ovs_interfaceid": "b739a840-515d-46e6-b889-5e9f8ec9ad71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.086232] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062208, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.113968] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52743964-164d-7dbc-e3ce-d3f280a07871, 'name': SearchDatastore_Task, 'duration_secs': 0.013397} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.114385] env[61987]: DEBUG oslo_concurrency.lockutils [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.114640] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1024.114883] env[61987]: DEBUG oslo_concurrency.lockutils [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.115070] env[61987]: DEBUG oslo_concurrency.lockutils [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.115275] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1024.115548] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3ff8c6e6-0564-4921-a877-45e52b7ac94f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.121366] env[61987]: DEBUG nova.compute.manager [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1024.121780] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1024.123754] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9202689f-b2cc-42ff-a915-6c2370e95a62 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.127500] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1024.127723] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1024.128913] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bdddffc2-022f-4c72-a134-5a170cceabc9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.134211] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1024.134801] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-447d36be-acc9-4853-9e79-c08008f55897 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.140296] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.297s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.140296] env[61987]: DEBUG nova.compute.manager [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1024.141196] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1024.141196] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52d4baac-95f8-fcb1-d474-567dfa6f9e8e" [ 1024.141196] env[61987]: _type = "Task" [ 1024.141196] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.141446] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.451s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.142959] env[61987]: INFO nova.compute.claims [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1024.150694] env[61987]: DEBUG oslo_vmware.api [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 1024.150694] env[61987]: value = "task-1062210" [ 1024.150694] env[61987]: _type = "Task" [ 1024.150694] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.157570] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52d4baac-95f8-fcb1-d474-567dfa6f9e8e, 'name': SearchDatastore_Task, 'duration_secs': 0.008477} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.158768] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0bb3836-37b6-4db8-be71-db99401831c9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.164091] env[61987]: DEBUG oslo_vmware.api [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062210, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.167570] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1024.167570] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]522cb4a7-0e89-c0ae-bc9f-e5d4ab33a01e" [ 1024.167570] env[61987]: _type = "Task" [ 1024.167570] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.177119] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]522cb4a7-0e89-c0ae-bc9f-e5d4ab33a01e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.381271] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062209, 'name': PowerOffVM_Task, 'duration_secs': 0.203225} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.381642] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1024.382738] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2074b27-e2a1-4272-85a4-f7d3134bfe78 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.400755] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a0cc44f-51fe-4dbe-b615-50cc18b7655c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.403799] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.404120] env[61987]: DEBUG nova.compute.manager [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Instance network_info: |[{"id": "b739a840-515d-46e6-b889-5e9f8ec9ad71", "address": "fa:16:3e:cb:9f:9b", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb739a840-51", "ovs_interfaceid": "b739a840-515d-46e6-b889-5e9f8ec9ad71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1024.404901] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cb:9f:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a1c8fdf9-970c-4ae0-b6d9-f1015196b552', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b739a840-515d-46e6-b889-5e9f8ec9ad71', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1024.412167] env[61987]: DEBUG oslo.service.loopingcall [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1024.412752] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1024.413092] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-03fd6c14-0fba-4f1b-8a65-06410858eaef {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.438458] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1024.438458] env[61987]: value = "task-1062211" [ 1024.438458] env[61987]: _type = "Task" [ 1024.438458] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.446815] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062211, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.454460] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1024.454755] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c1afd043-f07f-4b51-9a2b-1ef068d087a0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.460835] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1024.460835] env[61987]: value = "task-1062212" [ 1024.460835] env[61987]: _type = "Task" [ 1024.460835] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.470601] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062212, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.476606] env[61987]: DEBUG nova.compute.manager [req-2d603af3-1309-497e-8724-d61f11da98c8 req-53644791-f44d-4a7f-a201-bcca61f0aa72 service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Received event network-changed-b739a840-515d-46e6-b889-5e9f8ec9ad71 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1024.476812] env[61987]: DEBUG nova.compute.manager [req-2d603af3-1309-497e-8724-d61f11da98c8 req-53644791-f44d-4a7f-a201-bcca61f0aa72 service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Refreshing instance network info cache due to event network-changed-b739a840-515d-46e6-b889-5e9f8ec9ad71. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1024.477039] env[61987]: DEBUG oslo_concurrency.lockutils [req-2d603af3-1309-497e-8724-d61f11da98c8 req-53644791-f44d-4a7f-a201-bcca61f0aa72 service nova] Acquiring lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.477194] env[61987]: DEBUG oslo_concurrency.lockutils [req-2d603af3-1309-497e-8724-d61f11da98c8 req-53644791-f44d-4a7f-a201-bcca61f0aa72 service nova] Acquired lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.477725] env[61987]: DEBUG nova.network.neutron [req-2d603af3-1309-497e-8724-d61f11da98c8 req-53644791-f44d-4a7f-a201-bcca61f0aa72 service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Refreshing network info cache for port b739a840-515d-46e6-b889-5e9f8ec9ad71 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1024.586022] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062208, 'name': ReconfigVM_Task, 'duration_secs': 0.560266} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.586364] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 0da22c35-f11d-4498-bdb4-31bfd4c364e6/0da22c35-f11d-4498-bdb4-31bfd4c364e6.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1024.586976] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3047aaa4-e710-4d01-a12c-d90a007da615 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.592833] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for the task: (returnval){ [ 1024.592833] env[61987]: value = "task-1062213" [ 1024.592833] env[61987]: _type = "Task" [ 1024.592833] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.600435] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062213, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.650356] env[61987]: DEBUG nova.compute.utils [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1024.653426] env[61987]: DEBUG nova.compute.manager [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1024.653426] env[61987]: DEBUG nova.network.neutron [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1024.664197] env[61987]: DEBUG oslo_vmware.api [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062210, 'name': PowerOffVM_Task, 'duration_secs': 0.17756} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.665481] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1024.665947] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1024.666207] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c6ae8e13-6dd0-4921-bb43-e237224dae5a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.678150] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]522cb4a7-0e89-c0ae-bc9f-e5d4ab33a01e, 'name': SearchDatastore_Task, 'duration_secs': 0.011779} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.678418] env[61987]: DEBUG oslo_concurrency.lockutils [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.678682] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] b172195a-79d2-4f24-b9df-320a07de965e/b172195a-79d2-4f24-b9df-320a07de965e.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1024.678935] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e9f452ef-1ebe-450e-9c97-33d70e6dfd65 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.684901] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1024.684901] env[61987]: value = "task-1062215" [ 1024.684901] env[61987]: _type = "Task" [ 1024.684901] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.692567] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062215, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.728217] env[61987]: DEBUG nova.policy [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0945d6196b8c473da5308e6d784ab479', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b77fbbfc7f6437083f9f30e8fdeeb35', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 1024.733977] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1024.734295] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1024.734489] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Deleting the datastore file [datastore1] 84a4a69b-04db-4546-ba89-9039e382a0c4 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1024.734777] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b670a3fc-5135-46df-9203-57b7ade8e541 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.741726] env[61987]: DEBUG oslo_vmware.api [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for the task: (returnval){ [ 1024.741726] env[61987]: value = "task-1062216" [ 1024.741726] env[61987]: _type = "Task" [ 1024.741726] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.755199] env[61987]: DEBUG oslo_vmware.api [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062216, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.841421] env[61987]: INFO nova.compute.manager [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Rebuilding instance [ 1024.898928] env[61987]: DEBUG nova.compute.manager [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1024.899857] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82e07418-5c56-4c2d-a835-64b52c924410 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.949145] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062211, 'name': CreateVM_Task, 'duration_secs': 0.395746} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.949375] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1024.950485] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.950485] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.950747] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1024.951069] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac7525cd-b6a7-4011-9d05-dc83f2f7dbf7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.957032] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1024.957032] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52305259-19b4-f598-8fa1-f5bbb68a21c4" [ 1024.957032] env[61987]: _type = "Task" [ 1024.957032] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.974339] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52305259-19b4-f598-8fa1-f5bbb68a21c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.978212] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] VM already powered off {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1024.978212] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1024.978212] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.978212] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.978212] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1024.978696] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f941c5e5-0106-4e4f-804d-781f825e892c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.991022] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1024.991022] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1024.991895] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4fc01d4f-ef49-4e21-adef-56b6043ef642 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.999520] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1024.999520] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5290bf0d-bcb1-4934-ee5b-5e13a6372209" [ 1024.999520] env[61987]: _type = "Task" [ 1024.999520] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.012157] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5290bf0d-bcb1-4934-ee5b-5e13a6372209, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.104954] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062213, 'name': Rename_Task, 'duration_secs': 0.161669} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.108470] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1025.108739] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f25acfb0-228a-40b6-948a-6bb147dd1523 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.117269] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for the task: (returnval){ [ 1025.117269] env[61987]: value = "task-1062217" [ 1025.117269] env[61987]: _type = "Task" [ 1025.117269] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.126916] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062217, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.156801] env[61987]: DEBUG nova.compute.manager [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1025.196364] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062215, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.257556] env[61987]: DEBUG nova.network.neutron [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Successfully created port: a46b1587-a8f9-4552-8d05-57882397c3f5 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1025.265068] env[61987]: DEBUG oslo_vmware.api [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Task: {'id': task-1062216, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.195995} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.265455] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1025.265455] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1025.265634] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1025.266317] env[61987]: INFO nova.compute.manager [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1025.266317] env[61987]: DEBUG oslo.service.loopingcall [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1025.266317] env[61987]: DEBUG nova.compute.manager [-] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1025.266432] env[61987]: DEBUG nova.network.neutron [-] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1025.430723] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d9ce69-b547-4c30-84c8-df39519ea192 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.437981] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87dbb6e6-d78f-4001-b5e6-a96a2277f32c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.483051] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cef1374f-f079-460a-b566-ce5ff7cf613b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.492176] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a99f1243-48ee-4dbf-b9d9-a6a9e395edea {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.497093] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52305259-19b4-f598-8fa1-f5bbb68a21c4, 'name': SearchDatastore_Task, 'duration_secs': 0.072255} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.497513] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.497806] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1025.498133] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.498293] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.498524] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1025.499241] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f33ad562-9383-463d-9b62-69743d51fb49 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.510507] env[61987]: DEBUG nova.compute.provider_tree [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.522498] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5290bf0d-bcb1-4934-ee5b-5e13a6372209, 'name': SearchDatastore_Task, 'duration_secs': 0.021869} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.524449] env[61987]: DEBUG nova.network.neutron [req-2d603af3-1309-497e-8724-d61f11da98c8 req-53644791-f44d-4a7f-a201-bcca61f0aa72 service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Updated VIF entry in instance network info cache for port b739a840-515d-46e6-b889-5e9f8ec9ad71. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1025.525300] env[61987]: DEBUG nova.network.neutron [req-2d603af3-1309-497e-8724-d61f11da98c8 req-53644791-f44d-4a7f-a201-bcca61f0aa72 service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Updating instance_info_cache with network_info: [{"id": "b739a840-515d-46e6-b889-5e9f8ec9ad71", "address": "fa:16:3e:cb:9f:9b", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb739a840-51", "ovs_interfaceid": "b739a840-515d-46e6-b889-5e9f8ec9ad71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.527686] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1025.527937] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1025.530188] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ef674aa-7947-4b5f-803b-93c2e8780cce {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.533349] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73f982f0-b000-4c7e-82af-7806dc821de4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.541508] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1025.541508] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5259f3d7-0f00-f95a-4858-b23ae7a386a1" [ 1025.541508] env[61987]: _type = "Task" [ 1025.541508] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.542958] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1025.542958] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5212cb50-0430-63f8-1cea-fd0aa66a83bf" [ 1025.542958] env[61987]: _type = "Task" [ 1025.542958] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.554661] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5259f3d7-0f00-f95a-4858-b23ae7a386a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.558420] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5212cb50-0430-63f8-1cea-fd0aa66a83bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.627776] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062217, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.698191] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062215, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519411} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.698463] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] b172195a-79d2-4f24-b9df-320a07de965e/b172195a-79d2-4f24-b9df-320a07de965e.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1025.698719] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1025.698950] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-40aad91d-f0b0-444f-b448-7c26ab22d1ce {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.705977] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1025.705977] env[61987]: value = "task-1062218" [ 1025.705977] env[61987]: _type = "Task" [ 1025.705977] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.714357] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062218, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.915350] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1025.915742] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-48ba33e7-8861-4c8f-8a07-dbf0a6691f79 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.922918] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1025.922918] env[61987]: value = "task-1062219" [ 1025.922918] env[61987]: _type = "Task" [ 1025.922918] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.933320] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062219, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.013711] env[61987]: DEBUG nova.scheduler.client.report [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1026.030673] env[61987]: DEBUG oslo_concurrency.lockutils [req-2d603af3-1309-497e-8724-d61f11da98c8 req-53644791-f44d-4a7f-a201-bcca61f0aa72 service nova] Releasing lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.055757] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5259f3d7-0f00-f95a-4858-b23ae7a386a1, 'name': SearchDatastore_Task, 'duration_secs': 0.013858} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.059164] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.059541] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] dab6d2c5-8c04-4977-a409-b384037eabce/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk. {{(pid=61987) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1026.059865] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5212cb50-0430-63f8-1cea-fd0aa66a83bf, 'name': SearchDatastore_Task, 'duration_secs': 0.012735} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.060122] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-30877367-e031-4e7f-b730-869b8cfec298 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.062804] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4992f42-4c11-43fe-a4a8-f89fc74526b5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.069213] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1026.069213] env[61987]: value = "task-1062220" [ 1026.069213] env[61987]: _type = "Task" [ 1026.069213] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.070361] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1026.070361] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52214ab6-ccd8-f636-6d96-24945f81ff5b" [ 1026.070361] env[61987]: _type = "Task" [ 1026.070361] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.080971] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062220, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.083614] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52214ab6-ccd8-f636-6d96-24945f81ff5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.128041] env[61987]: DEBUG oslo_vmware.api [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062217, 'name': PowerOnVM_Task, 'duration_secs': 0.639078} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.128415] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1026.128698] env[61987]: DEBUG nova.compute.manager [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1026.129617] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d40cb7-19e5-491d-a09c-1ac5e1521209 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.168600] env[61987]: DEBUG nova.compute.manager [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1026.194588] env[61987]: DEBUG nova.virt.hardware [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1026.194864] env[61987]: DEBUG nova.virt.hardware [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1026.195095] env[61987]: DEBUG nova.virt.hardware [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1026.195313] env[61987]: DEBUG nova.virt.hardware [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1026.195470] env[61987]: DEBUG nova.virt.hardware [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1026.195624] env[61987]: DEBUG nova.virt.hardware [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1026.195861] env[61987]: DEBUG nova.virt.hardware [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1026.196059] env[61987]: DEBUG nova.virt.hardware [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1026.196279] env[61987]: DEBUG nova.virt.hardware [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1026.196472] env[61987]: DEBUG nova.virt.hardware [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1026.196675] env[61987]: DEBUG nova.virt.hardware [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1026.197643] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a0beb45-940e-4c25-9cb0-e951e327077b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.205837] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb37d685-d7d3-498e-bc58-036cac14413d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.226815] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062218, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.162887} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.227127] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1026.227954] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf19bdd-0013-4c11-a388-2cc762e07137 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.254485] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] b172195a-79d2-4f24-b9df-320a07de965e/b172195a-79d2-4f24-b9df-320a07de965e.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1026.254895] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97a97988-ed92-4a71-b759-9100a0f58375 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.276915] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1026.276915] env[61987]: value = "task-1062221" [ 1026.276915] env[61987]: _type = "Task" [ 1026.276915] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.289040] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062221, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.369435] env[61987]: DEBUG nova.network.neutron [-] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.442378] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062219, 'name': PowerOffVM_Task, 'duration_secs': 0.381777} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.442866] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1026.443320] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1026.444562] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9ba6c70-e568-4de6-84fa-c8ef7e3dd0b5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.454486] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1026.455606] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b0a89102-3b29-4816-b03f-a56821fc734c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.513218] env[61987]: DEBUG nova.compute.manager [req-c0899140-fae2-48a1-9528-272801614c5a req-df63264b-39c2-4609-8ebf-d764bd42f42d service nova] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Received event network-vif-deleted-c9f4e2e0-50b8-4e72-a06d-8c2729b200ed {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1026.519150] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.377s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.519938] env[61987]: DEBUG nova.compute.manager [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1026.533615] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1026.533851] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1026.534158] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Deleting the datastore file [datastore1] f3773cda-41e8-4804-9319-c7e458112ddf {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1026.534459] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fa8070b4-cf57-4684-bdd3-f0e7ccaf18ec {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.542110] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1026.542110] env[61987]: value = "task-1062223" [ 1026.542110] env[61987]: _type = "Task" [ 1026.542110] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.552789] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062223, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.584034] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062220, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.480074} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.587628] env[61987]: INFO nova.virt.vmwareapi.ds_util [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] dab6d2c5-8c04-4977-a409-b384037eabce/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk. [ 1026.587998] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52214ab6-ccd8-f636-6d96-24945f81ff5b, 'name': SearchDatastore_Task, 'duration_secs': 0.012007} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.588727] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-679af327-6d26-4ae6-a1c1-15bc65de6481 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.591296] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.591597] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 0cba5e51-7e0a-409b-a7b2-612ab8cf5689/0cba5e51-7e0a-409b-a7b2-612ab8cf5689.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1026.591849] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f9986477-be08-4fad-9659-e9a9d0ace68d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.618116] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] dab6d2c5-8c04-4977-a409-b384037eabce/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1026.619703] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f40441b6-7f00-4c67-a8ac-380e0cd9d285 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.632795] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1026.632795] env[61987]: value = "task-1062224" [ 1026.632795] env[61987]: _type = "Task" [ 1026.632795] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.638554] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1026.638554] env[61987]: value = "task-1062225" [ 1026.638554] env[61987]: _type = "Task" [ 1026.638554] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.645112] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062224, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.651217] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.651309] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.651510] env[61987]: DEBUG nova.objects.instance [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61987) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1026.659645] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062225, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.788951] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062221, 'name': ReconfigVM_Task, 'duration_secs': 0.34365} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.789276] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Reconfigured VM instance instance-0000006b to attach disk [datastore2] b172195a-79d2-4f24-b9df-320a07de965e/b172195a-79d2-4f24-b9df-320a07de965e.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1026.789976] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-88ea2a44-0fd8-4002-a368-d57de9d38f3b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.797700] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1026.797700] env[61987]: value = "task-1062226" [ 1026.797700] env[61987]: _type = "Task" [ 1026.797700] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.807202] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062226, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.873657] env[61987]: INFO nova.compute.manager [-] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Took 1.61 seconds to deallocate network for instance. [ 1027.026384] env[61987]: DEBUG nova.compute.utils [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1027.027965] env[61987]: DEBUG nova.compute.manager [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1027.028222] env[61987]: DEBUG nova.network.neutron [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1027.054040] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062223, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176098} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.054390] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1027.054586] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1027.054768] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1027.105915] env[61987]: DEBUG nova.policy [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '914fdab15c68437db171f9710ea8ae85', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3a234a62da0e498fbe613fbcaaec3201', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 1027.142925] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062224, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.501055} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.146562] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 0cba5e51-7e0a-409b-a7b2-612ab8cf5689/0cba5e51-7e0a-409b-a7b2-612ab8cf5689.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1027.146793] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1027.147054] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8799d631-7ef5-4968-b67d-8603497438c2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.151417] env[61987]: DEBUG nova.compute.manager [req-1895ca0d-eabf-4bb4-837c-b565f445a91a req-249be62f-add3-4c0f-b4e6-8b348ae6df82 service nova] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Received event network-vif-plugged-a46b1587-a8f9-4552-8d05-57882397c3f5 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1027.151627] env[61987]: DEBUG oslo_concurrency.lockutils [req-1895ca0d-eabf-4bb4-837c-b565f445a91a req-249be62f-add3-4c0f-b4e6-8b348ae6df82 service nova] Acquiring lock "bf1123bf-80e6-4be3-804f-e13b906ee44d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.151837] env[61987]: DEBUG oslo_concurrency.lockutils [req-1895ca0d-eabf-4bb4-837c-b565f445a91a req-249be62f-add3-4c0f-b4e6-8b348ae6df82 service nova] Lock "bf1123bf-80e6-4be3-804f-e13b906ee44d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.152029] env[61987]: DEBUG oslo_concurrency.lockutils [req-1895ca0d-eabf-4bb4-837c-b565f445a91a req-249be62f-add3-4c0f-b4e6-8b348ae6df82 service nova] Lock "bf1123bf-80e6-4be3-804f-e13b906ee44d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.152211] env[61987]: DEBUG nova.compute.manager [req-1895ca0d-eabf-4bb4-837c-b565f445a91a req-249be62f-add3-4c0f-b4e6-8b348ae6df82 service nova] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] No waiting events found dispatching network-vif-plugged-a46b1587-a8f9-4552-8d05-57882397c3f5 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1027.152396] env[61987]: WARNING nova.compute.manager [req-1895ca0d-eabf-4bb4-837c-b565f445a91a req-249be62f-add3-4c0f-b4e6-8b348ae6df82 service nova] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Received unexpected event network-vif-plugged-a46b1587-a8f9-4552-8d05-57882397c3f5 for instance with vm_state building and task_state spawning. [ 1027.159470] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062225, 'name': ReconfigVM_Task, 'duration_secs': 0.315446} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.159749] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1027.159749] env[61987]: value = "task-1062227" [ 1027.159749] env[61987]: _type = "Task" [ 1027.159749] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.160559] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Reconfigured VM instance instance-00000069 to attach disk [datastore1] dab6d2c5-8c04-4977-a409-b384037eabce/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1027.161475] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e63b27-f490-4725-be46-ebb08c8875b7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.196562] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9420b5c1-b883-4c8b-85b4-8587bd2c024f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.207287] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062227, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.212829] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1027.212829] env[61987]: value = "task-1062228" [ 1027.212829] env[61987]: _type = "Task" [ 1027.212829] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.220810] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062228, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.308335] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062226, 'name': Rename_Task, 'duration_secs': 0.224865} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.309308] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1027.309308] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f7d1dc7f-87cb-4254-afb0-5a07a18eb049 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.315121] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1027.315121] env[61987]: value = "task-1062229" [ 1027.315121] env[61987]: _type = "Task" [ 1027.315121] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.328943] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062229, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.380336] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.457652] env[61987]: DEBUG nova.network.neutron [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Successfully created port: d96dede6-c439-4a14-a4d1-62c44e03ea11 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1027.533843] env[61987]: DEBUG nova.compute.manager [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1027.664147] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0ce7d993-40fe-47c7-aa52-077f1e31281f tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.664147] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.283s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.664147] env[61987]: DEBUG nova.objects.instance [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lazy-loading 'resources' on Instance uuid 84a4a69b-04db-4546-ba89-9039e382a0c4 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.673692] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062227, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063698} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.673980] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1027.674842] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84ff24ea-0489-44ea-a2ec-01125fc20554 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.700305] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] 0cba5e51-7e0a-409b-a7b2-612ab8cf5689/0cba5e51-7e0a-409b-a7b2-612ab8cf5689.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1027.700873] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-913f2b99-f4ff-4414-a904-5ef3c4445ec3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.732597] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1027.732597] env[61987]: value = "task-1062230" [ 1027.732597] env[61987]: _type = "Task" [ 1027.732597] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.733266] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062228, 'name': ReconfigVM_Task, 'duration_secs': 0.165125} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.733626] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1027.737187] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c18183ba-38db-4cef-94d1-614945ac9315 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.745444] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062230, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.747326] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1027.747326] env[61987]: value = "task-1062231" [ 1027.747326] env[61987]: _type = "Task" [ 1027.747326] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.748562] env[61987]: DEBUG nova.network.neutron [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Successfully updated port: a46b1587-a8f9-4552-8d05-57882397c3f5 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1027.763727] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062231, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.824986] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062229, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.833670] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Acquiring lock "0da22c35-f11d-4498-bdb4-31bfd4c364e6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.833939] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Lock "0da22c35-f11d-4498-bdb4-31bfd4c364e6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.834182] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Acquiring lock "0da22c35-f11d-4498-bdb4-31bfd4c364e6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.834378] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Lock "0da22c35-f11d-4498-bdb4-31bfd4c364e6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.834566] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Lock "0da22c35-f11d-4498-bdb4-31bfd4c364e6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.836949] env[61987]: INFO nova.compute.manager [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Terminating instance [ 1028.090718] env[61987]: DEBUG nova.virt.hardware [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1028.090995] env[61987]: DEBUG nova.virt.hardware [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1028.091180] env[61987]: DEBUG nova.virt.hardware [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1028.091378] env[61987]: DEBUG nova.virt.hardware [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1028.091532] env[61987]: DEBUG nova.virt.hardware [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1028.091681] env[61987]: DEBUG nova.virt.hardware [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1028.091894] env[61987]: DEBUG nova.virt.hardware [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1028.092079] env[61987]: DEBUG nova.virt.hardware [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1028.092308] env[61987]: DEBUG nova.virt.hardware [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1028.092496] env[61987]: DEBUG nova.virt.hardware [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1028.092722] env[61987]: DEBUG nova.virt.hardware [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1028.093962] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a82e186-ca25-402a-aa20-e0b337cbd688 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.102723] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b0f605-55e4-40a0-bf06-5cd554dcb772 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.116615] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:00:34:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2a111da1-447a-4722-9662-b0d6a2886a65', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b5999bb0-480a-45bd-b07c-fbb91985421e', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1028.124321] env[61987]: DEBUG oslo.service.loopingcall [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1028.124792] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1028.125057] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-274b9f5d-784e-43fd-8ce4-f32b16893de5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.144307] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1028.144307] env[61987]: value = "task-1062232" [ 1028.144307] env[61987]: _type = "Task" [ 1028.144307] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.154342] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062232, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.243588] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062230, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.255271] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "refresh_cache-bf1123bf-80e6-4be3-804f-e13b906ee44d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.255580] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquired lock "refresh_cache-bf1123bf-80e6-4be3-804f-e13b906ee44d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.255681] env[61987]: DEBUG nova.network.neutron [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1028.263348] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062231, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.295078] env[61987]: DEBUG nova.network.neutron [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1028.328983] env[61987]: DEBUG oslo_vmware.api [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062229, 'name': PowerOnVM_Task, 'duration_secs': 0.703954} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.329284] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1028.329508] env[61987]: INFO nova.compute.manager [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Took 8.55 seconds to spawn the instance on the hypervisor. [ 1028.329692] env[61987]: DEBUG nova.compute.manager [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1028.330509] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76012bc5-8823-438c-b7f8-4b298e32d97d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.344063] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Acquiring lock "refresh_cache-0da22c35-f11d-4498-bdb4-31bfd4c364e6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.344063] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Acquired lock "refresh_cache-0da22c35-f11d-4498-bdb4-31bfd4c364e6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.344063] env[61987]: DEBUG nova.network.neutron [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1028.388713] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09369fc1-1da6-4ced-b674-5580980637f3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.397241] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f8b0db2-21e7-4ee2-ac19-d3072f9ccb1e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.429014] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5924cfa-e366-4ea1-9b22-2f9db3f162a9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.436801] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0265b3-960e-4d6b-86be-081325a6ca84 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.451849] env[61987]: DEBUG nova.compute.provider_tree [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.475721] env[61987]: DEBUG nova.network.neutron [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Updating instance_info_cache with network_info: [{"id": "a46b1587-a8f9-4552-8d05-57882397c3f5", "address": "fa:16:3e:b2:84:05", "network": {"id": "26d453d1-8edf-45e1-9e49-2a3225de6633", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1400564876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b77fbbfc7f6437083f9f30e8fdeeb35", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa46b1587-a8", "ovs_interfaceid": "a46b1587-a8f9-4552-8d05-57882397c3f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.539222] env[61987]: DEBUG nova.compute.manager [req-4bad1335-faa0-43cf-82dc-53a5ca08434f req-763a46fc-8f7a-475d-93e4-61641fc22f0d service nova] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Received event network-changed-a46b1587-a8f9-4552-8d05-57882397c3f5 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1028.539403] env[61987]: DEBUG nova.compute.manager [req-4bad1335-faa0-43cf-82dc-53a5ca08434f req-763a46fc-8f7a-475d-93e4-61641fc22f0d service nova] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Refreshing instance network info cache due to event network-changed-a46b1587-a8f9-4552-8d05-57882397c3f5. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1028.539581] env[61987]: DEBUG oslo_concurrency.lockutils [req-4bad1335-faa0-43cf-82dc-53a5ca08434f req-763a46fc-8f7a-475d-93e4-61641fc22f0d service nova] Acquiring lock "refresh_cache-bf1123bf-80e6-4be3-804f-e13b906ee44d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.546019] env[61987]: DEBUG nova.compute.manager [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1028.569022] env[61987]: DEBUG nova.virt.hardware [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1028.569370] env[61987]: DEBUG nova.virt.hardware [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1028.569552] env[61987]: DEBUG nova.virt.hardware [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1028.569758] env[61987]: DEBUG nova.virt.hardware [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1028.569924] env[61987]: DEBUG nova.virt.hardware [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1028.570111] env[61987]: DEBUG nova.virt.hardware [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1028.570535] env[61987]: DEBUG nova.virt.hardware [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1028.570535] env[61987]: DEBUG nova.virt.hardware [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1028.570684] env[61987]: DEBUG nova.virt.hardware [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1028.570864] env[61987]: DEBUG nova.virt.hardware [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1028.571073] env[61987]: DEBUG nova.virt.hardware [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1028.572167] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcc58631-7b6d-4bb3-865c-5f2523e1a43c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.580791] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f0dd274-de66-4bc4-add1-7b0503b80379 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.654160] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062232, 'name': CreateVM_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.744443] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062230, 'name': ReconfigVM_Task, 'duration_secs': 0.629816} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.746043] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Reconfigured VM instance instance-0000006c to attach disk [datastore2] 0cba5e51-7e0a-409b-a7b2-612ab8cf5689/0cba5e51-7e0a-409b-a7b2-612ab8cf5689.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1028.746179] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-65b0f4e1-be0e-4890-b110-988b4c6b648c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.755615] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1028.755615] env[61987]: value = "task-1062233" [ 1028.755615] env[61987]: _type = "Task" [ 1028.755615] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.762258] env[61987]: DEBUG oslo_vmware.api [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062231, 'name': PowerOnVM_Task, 'duration_secs': 0.537133} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.762962] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1028.765988] env[61987]: DEBUG nova.compute.manager [None req-b83f6f27-1ec1-4e0a-8d74-b000b80a1495 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1028.769822] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf16821-6657-4a9e-8292-7c85fb8d7bb8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.772527] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062233, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.855621] env[61987]: INFO nova.compute.manager [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Took 16.11 seconds to build instance. [ 1028.882867] env[61987]: DEBUG nova.network.neutron [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1028.955218] env[61987]: DEBUG nova.scheduler.client.report [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1028.977792] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Releasing lock "refresh_cache-bf1123bf-80e6-4be3-804f-e13b906ee44d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.978118] env[61987]: DEBUG nova.compute.manager [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Instance network_info: |[{"id": "a46b1587-a8f9-4552-8d05-57882397c3f5", "address": "fa:16:3e:b2:84:05", "network": {"id": "26d453d1-8edf-45e1-9e49-2a3225de6633", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1400564876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b77fbbfc7f6437083f9f30e8fdeeb35", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa46b1587-a8", "ovs_interfaceid": "a46b1587-a8f9-4552-8d05-57882397c3f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1028.978440] env[61987]: DEBUG oslo_concurrency.lockutils [req-4bad1335-faa0-43cf-82dc-53a5ca08434f req-763a46fc-8f7a-475d-93e4-61641fc22f0d service nova] Acquired lock "refresh_cache-bf1123bf-80e6-4be3-804f-e13b906ee44d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.978629] env[61987]: DEBUG nova.network.neutron [req-4bad1335-faa0-43cf-82dc-53a5ca08434f req-763a46fc-8f7a-475d-93e4-61641fc22f0d service nova] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Refreshing network info cache for port a46b1587-a8f9-4552-8d05-57882397c3f5 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1028.980302] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b2:84:05', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3734b156-0f7d-4721-b23c-d000412ec2eb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a46b1587-a8f9-4552-8d05-57882397c3f5', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1028.988272] env[61987]: DEBUG oslo.service.loopingcall [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1028.989140] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1028.989389] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-139cad7c-59f0-4477-a9bd-be516245a704 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.006996] env[61987]: DEBUG nova.network.neutron [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.010268] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1029.010268] env[61987]: value = "task-1062234" [ 1029.010268] env[61987]: _type = "Task" [ 1029.010268] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.023858] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062234, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.157518] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062232, 'name': CreateVM_Task, 'duration_secs': 0.70232} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.157697] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1029.158374] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.158549] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.158867] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1029.159329] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7ec2e3c-58bb-48bc-bebd-6ff78c90ac1f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.164374] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1029.164374] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5226d0b4-5424-8350-fc79-c2877990e7e6" [ 1029.164374] env[61987]: _type = "Task" [ 1029.164374] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.174137] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5226d0b4-5424-8350-fc79-c2877990e7e6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.269451] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062233, 'name': Rename_Task, 'duration_secs': 0.368059} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.269816] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1029.270089] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b4430cf1-ef7e-4274-8d4c-a0c763f55feb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.277339] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1029.277339] env[61987]: value = "task-1062235" [ 1029.277339] env[61987]: _type = "Task" [ 1029.277339] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.283860] env[61987]: DEBUG nova.network.neutron [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Successfully updated port: d96dede6-c439-4a14-a4d1-62c44e03ea11 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1029.293798] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062235, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.357701] env[61987]: DEBUG oslo_concurrency.lockutils [None req-27bdfb8f-4f84-4498-b539-8206d497dd59 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "b172195a-79d2-4f24-b9df-320a07de965e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.618s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.460090] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.796s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.484118] env[61987]: INFO nova.scheduler.client.report [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Deleted allocations for instance 84a4a69b-04db-4546-ba89-9039e382a0c4 [ 1029.510220] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Releasing lock "refresh_cache-0da22c35-f11d-4498-bdb4-31bfd4c364e6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.510836] env[61987]: DEBUG nova.compute.manager [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1029.510964] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1029.511910] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba9caa26-e209-46d3-aa9b-55651a01f2c4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.526010] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062234, 'name': CreateVM_Task, 'duration_secs': 0.491794} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.528371] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1029.528371] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1029.528855] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.529061] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3013e07e-46d6-4c91-8184-68db8c8ec386 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.536115] env[61987]: DEBUG oslo_vmware.api [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for the task: (returnval){ [ 1029.536115] env[61987]: value = "task-1062236" [ 1029.536115] env[61987]: _type = "Task" [ 1029.536115] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.542794] env[61987]: DEBUG oslo_vmware.api [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062236, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.621812] env[61987]: DEBUG oslo_concurrency.lockutils [None req-71ba2865-5169-40f4-8456-f2f4d5d69ecd tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "b172195a-79d2-4f24-b9df-320a07de965e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.622127] env[61987]: DEBUG oslo_concurrency.lockutils [None req-71ba2865-5169-40f4-8456-f2f4d5d69ecd tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "b172195a-79d2-4f24-b9df-320a07de965e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.622328] env[61987]: DEBUG nova.compute.manager [None req-71ba2865-5169-40f4-8456-f2f4d5d69ecd tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1029.626447] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5581835b-eaba-4470-9ba3-5a848f7e6eeb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.633693] env[61987]: DEBUG nova.compute.manager [None req-71ba2865-5169-40f4-8456-f2f4d5d69ecd tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61987) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1029.634382] env[61987]: DEBUG nova.objects.instance [None req-71ba2865-5169-40f4-8456-f2f4d5d69ecd tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lazy-loading 'flavor' on Instance uuid b172195a-79d2-4f24-b9df-320a07de965e {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1029.675652] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5226d0b4-5424-8350-fc79-c2877990e7e6, 'name': SearchDatastore_Task, 'duration_secs': 0.010345} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.678187] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.678445] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1029.678688] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.678858] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.679044] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1029.679351] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.679662] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1029.679905] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3f42dc53-6799-45cc-9e80-5cad4543971d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.682011] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-edce1f3c-32b8-4490-8263-1051447de777 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.688959] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 1029.688959] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]524c3e4b-c201-52e1-6283-022cc93e0f81" [ 1029.688959] env[61987]: _type = "Task" [ 1029.688959] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.690082] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1029.690271] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1029.693493] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49f8bddf-0f6e-4ee7-b180-eca7e0054a27 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.701577] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1029.701577] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5226e36b-ac13-bede-e0ac-02675b6a8f83" [ 1029.701577] env[61987]: _type = "Task" [ 1029.701577] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.701791] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524c3e4b-c201-52e1-6283-022cc93e0f81, 'name': SearchDatastore_Task, 'duration_secs': 0.009634} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.702121] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.702357] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1029.702568] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.710063] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5226e36b-ac13-bede-e0ac-02675b6a8f83, 'name': SearchDatastore_Task, 'duration_secs': 0.008175} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.710778] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9fb31f0-4f31-46cc-93ee-743daea90e9d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.715689] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1029.715689] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52770284-7cc6-bebf-a19c-c282d6f69e44" [ 1029.715689] env[61987]: _type = "Task" [ 1029.715689] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.723504] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52770284-7cc6-bebf-a19c-c282d6f69e44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.737312] env[61987]: DEBUG nova.network.neutron [req-4bad1335-faa0-43cf-82dc-53a5ca08434f req-763a46fc-8f7a-475d-93e4-61641fc22f0d service nova] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Updated VIF entry in instance network info cache for port a46b1587-a8f9-4552-8d05-57882397c3f5. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1029.737681] env[61987]: DEBUG nova.network.neutron [req-4bad1335-faa0-43cf-82dc-53a5ca08434f req-763a46fc-8f7a-475d-93e4-61641fc22f0d service nova] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Updating instance_info_cache with network_info: [{"id": "a46b1587-a8f9-4552-8d05-57882397c3f5", "address": "fa:16:3e:b2:84:05", "network": {"id": "26d453d1-8edf-45e1-9e49-2a3225de6633", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1400564876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b77fbbfc7f6437083f9f30e8fdeeb35", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa46b1587-a8", "ovs_interfaceid": "a46b1587-a8f9-4552-8d05-57882397c3f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.788123] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "refresh_cache-7002962c-76ed-4a8d-be2a-b2b118847fea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.788523] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquired lock "refresh_cache-7002962c-76ed-4a8d-be2a-b2b118847fea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.788523] env[61987]: DEBUG nova.network.neutron [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1029.789579] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062235, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.807849] env[61987]: INFO nova.compute.manager [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Unrescuing [ 1029.808088] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquiring lock "refresh_cache-dab6d2c5-8c04-4977-a409-b384037eabce" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.808254] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquired lock "refresh_cache-dab6d2c5-8c04-4977-a409-b384037eabce" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.808428] env[61987]: DEBUG nova.network.neutron [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1029.995463] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b717c747-67a6-49dd-bbcd-efea2c9cee9e tempest-ServersTestJSON-964537519 tempest-ServersTestJSON-964537519-project-member] Lock "84a4a69b-04db-4546-ba89-9039e382a0c4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.381s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.046057] env[61987]: DEBUG oslo_vmware.api [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062236, 'name': PowerOffVM_Task, 'duration_secs': 0.13001} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.046414] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1030.046597] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1030.046858] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e85fb7bc-2bbf-47f2-82e2-d1146b4c9d94 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.069810] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1030.070106] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1030.070310] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Deleting the datastore file [datastore2] 0da22c35-f11d-4498-bdb4-31bfd4c364e6 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1030.070621] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f3ad50da-3d5f-44cc-80cc-8b8ebe3e895e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.077611] env[61987]: DEBUG oslo_vmware.api [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for the task: (returnval){ [ 1030.077611] env[61987]: value = "task-1062238" [ 1030.077611] env[61987]: _type = "Task" [ 1030.077611] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.086479] env[61987]: DEBUG oslo_vmware.api [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062238, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.227171] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52770284-7cc6-bebf-a19c-c282d6f69e44, 'name': SearchDatastore_Task, 'duration_secs': 0.009647} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.227459] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.227730] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] f3773cda-41e8-4804-9319-c7e458112ddf/f3773cda-41e8-4804-9319-c7e458112ddf.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1030.228031] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.228232] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1030.228458] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3e15abef-ec64-46ab-b282-cd97405066c8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.230618] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-16679efe-eccd-4fee-881b-a773331c36af {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.236943] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1030.236943] env[61987]: value = "task-1062239" [ 1030.236943] env[61987]: _type = "Task" [ 1030.236943] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.241766] env[61987]: DEBUG oslo_concurrency.lockutils [req-4bad1335-faa0-43cf-82dc-53a5ca08434f req-763a46fc-8f7a-475d-93e4-61641fc22f0d service nova] Releasing lock "refresh_cache-bf1123bf-80e6-4be3-804f-e13b906ee44d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.242164] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1030.242348] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1030.243496] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20290c3b-0df6-4039-9ad4-5ea1d4d03064 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.248936] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062239, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.251952] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 1030.251952] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52c37e23-0740-31cd-0d95-68c0544a89fe" [ 1030.251952] env[61987]: _type = "Task" [ 1030.251952] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.260211] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c37e23-0740-31cd-0d95-68c0544a89fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.288716] env[61987]: DEBUG oslo_vmware.api [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062235, 'name': PowerOnVM_Task, 'duration_secs': 0.631242} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.289070] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1030.289316] env[61987]: INFO nova.compute.manager [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Took 8.20 seconds to spawn the instance on the hypervisor. [ 1030.289527] env[61987]: DEBUG nova.compute.manager [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1030.290375] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89665ff-b33e-438b-a85d-7f5f76753525 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.361979] env[61987]: DEBUG nova.network.neutron [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1030.593891] env[61987]: DEBUG oslo_vmware.api [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Task: {'id': task-1062238, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.224749} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.598139] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1030.598633] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1030.599031] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1030.599579] env[61987]: INFO nova.compute.manager [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Took 1.09 seconds to destroy the instance on the hypervisor. [ 1030.600062] env[61987]: DEBUG oslo.service.loopingcall [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1030.604999] env[61987]: DEBUG nova.compute.manager [-] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1030.604999] env[61987]: DEBUG nova.network.neutron [-] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1030.606872] env[61987]: DEBUG nova.compute.manager [req-cec50c54-bebe-44e6-bdfc-b23c21c0427e req-9982a308-0d1c-4f73-a271-c4bf6829c37d service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Received event network-vif-plugged-d96dede6-c439-4a14-a4d1-62c44e03ea11 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1030.607362] env[61987]: DEBUG oslo_concurrency.lockutils [req-cec50c54-bebe-44e6-bdfc-b23c21c0427e req-9982a308-0d1c-4f73-a271-c4bf6829c37d service nova] Acquiring lock "7002962c-76ed-4a8d-be2a-b2b118847fea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.607735] env[61987]: DEBUG oslo_concurrency.lockutils [req-cec50c54-bebe-44e6-bdfc-b23c21c0427e req-9982a308-0d1c-4f73-a271-c4bf6829c37d service nova] Lock "7002962c-76ed-4a8d-be2a-b2b118847fea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.608225] env[61987]: DEBUG oslo_concurrency.lockutils [req-cec50c54-bebe-44e6-bdfc-b23c21c0427e req-9982a308-0d1c-4f73-a271-c4bf6829c37d service nova] Lock "7002962c-76ed-4a8d-be2a-b2b118847fea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.608622] env[61987]: DEBUG nova.compute.manager [req-cec50c54-bebe-44e6-bdfc-b23c21c0427e req-9982a308-0d1c-4f73-a271-c4bf6829c37d service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] No waiting events found dispatching network-vif-plugged-d96dede6-c439-4a14-a4d1-62c44e03ea11 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1030.608931] env[61987]: WARNING nova.compute.manager [req-cec50c54-bebe-44e6-bdfc-b23c21c0427e req-9982a308-0d1c-4f73-a271-c4bf6829c37d service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Received unexpected event network-vif-plugged-d96dede6-c439-4a14-a4d1-62c44e03ea11 for instance with vm_state building and task_state spawning. [ 1030.609270] env[61987]: DEBUG nova.compute.manager [req-cec50c54-bebe-44e6-bdfc-b23c21c0427e req-9982a308-0d1c-4f73-a271-c4bf6829c37d service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Received event network-changed-d96dede6-c439-4a14-a4d1-62c44e03ea11 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1030.609582] env[61987]: DEBUG nova.compute.manager [req-cec50c54-bebe-44e6-bdfc-b23c21c0427e req-9982a308-0d1c-4f73-a271-c4bf6829c37d service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Refreshing instance network info cache due to event network-changed-d96dede6-c439-4a14-a4d1-62c44e03ea11. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1030.609916] env[61987]: DEBUG oslo_concurrency.lockutils [req-cec50c54-bebe-44e6-bdfc-b23c21c0427e req-9982a308-0d1c-4f73-a271-c4bf6829c37d service nova] Acquiring lock "refresh_cache-7002962c-76ed-4a8d-be2a-b2b118847fea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.640615] env[61987]: DEBUG nova.network.neutron [-] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1030.644109] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-71ba2865-5169-40f4-8456-f2f4d5d69ecd tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1030.644109] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-49598b5e-0483-439b-942f-6d701b79216d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.652659] env[61987]: DEBUG oslo_vmware.api [None req-71ba2865-5169-40f4-8456-f2f4d5d69ecd tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1030.652659] env[61987]: value = "task-1062240" [ 1030.652659] env[61987]: _type = "Task" [ 1030.652659] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.664286] env[61987]: DEBUG oslo_vmware.api [None req-71ba2865-5169-40f4-8456-f2f4d5d69ecd tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062240, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.719041] env[61987]: DEBUG nova.network.neutron [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Updating instance_info_cache with network_info: [{"id": "d96dede6-c439-4a14-a4d1-62c44e03ea11", "address": "fa:16:3e:d2:e0:b1", "network": {"id": "4ed4fbdc-05b0-44f8-a3d5-b7b288abd5e0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2067313775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a234a62da0e498fbe613fbcaaec3201", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd96dede6-c4", "ovs_interfaceid": "d96dede6-c439-4a14-a4d1-62c44e03ea11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.748106] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062239, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.758726] env[61987]: DEBUG nova.network.neutron [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Updating instance_info_cache with network_info: [{"id": "05cdb86d-67c4-4b34-9cd3-4e4b3ed04d58", "address": "fa:16:3e:51:95:76", "network": {"id": "1157ea83-0804-47d3-96b7-7a5a16809f1a", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-318208769-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "630439422a354de0b32ac5eefaa990b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69744f59-ecac-4b0b-831e-82a274d7acbb", "external-id": "nsx-vlan-transportzone-770", "segmentation_id": 770, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05cdb86d-67", "ovs_interfaceid": "05cdb86d-67c4-4b34-9cd3-4e4b3ed04d58", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.764055] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c37e23-0740-31cd-0d95-68c0544a89fe, 'name': SearchDatastore_Task, 'duration_secs': 0.008929} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.765544] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df2218ab-1434-4993-a0e5-3497f3b2e92f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.770785] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 1030.770785] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52c2e8e0-de37-a43b-43ab-67deb6783249" [ 1030.770785] env[61987]: _type = "Task" [ 1030.770785] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.779478] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c2e8e0-de37-a43b-43ab-67deb6783249, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.811147] env[61987]: INFO nova.compute.manager [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Took 18.00 seconds to build instance. [ 1031.142581] env[61987]: DEBUG nova.network.neutron [-] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.163910] env[61987]: DEBUG oslo_vmware.api [None req-71ba2865-5169-40f4-8456-f2f4d5d69ecd tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062240, 'name': PowerOffVM_Task, 'duration_secs': 0.335591} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.164246] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-71ba2865-5169-40f4-8456-f2f4d5d69ecd tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1031.164511] env[61987]: DEBUG nova.compute.manager [None req-71ba2865-5169-40f4-8456-f2f4d5d69ecd tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1031.165425] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f74b62-cac4-4fb9-b6d8-541967edc728 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.220642] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Releasing lock "refresh_cache-7002962c-76ed-4a8d-be2a-b2b118847fea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.221067] env[61987]: DEBUG nova.compute.manager [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Instance network_info: |[{"id": "d96dede6-c439-4a14-a4d1-62c44e03ea11", "address": "fa:16:3e:d2:e0:b1", "network": {"id": "4ed4fbdc-05b0-44f8-a3d5-b7b288abd5e0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2067313775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a234a62da0e498fbe613fbcaaec3201", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd96dede6-c4", "ovs_interfaceid": "d96dede6-c439-4a14-a4d1-62c44e03ea11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1031.221725] env[61987]: DEBUG oslo_concurrency.lockutils [req-cec50c54-bebe-44e6-bdfc-b23c21c0427e req-9982a308-0d1c-4f73-a271-c4bf6829c37d service nova] Acquired lock "refresh_cache-7002962c-76ed-4a8d-be2a-b2b118847fea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.221934] env[61987]: DEBUG nova.network.neutron [req-cec50c54-bebe-44e6-bdfc-b23c21c0427e req-9982a308-0d1c-4f73-a271-c4bf6829c37d service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Refreshing network info cache for port d96dede6-c439-4a14-a4d1-62c44e03ea11 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1031.223114] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d2:e0:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ddfb706a-add1-4e16-9ac4-d20b16a1df6d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd96dede6-c439-4a14-a4d1-62c44e03ea11', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1031.230704] env[61987]: DEBUG oslo.service.loopingcall [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1031.231457] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1031.231707] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d487ae57-2f90-4f72-958b-c333076f1c7c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.256302] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062239, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.515383} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.257463] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] f3773cda-41e8-4804-9319-c7e458112ddf/f3773cda-41e8-4804-9319-c7e458112ddf.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1031.257688] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1031.258132] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1031.258132] env[61987]: value = "task-1062241" [ 1031.258132] env[61987]: _type = "Task" [ 1031.258132] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.258379] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ec0e616c-9312-4dcd-8775-785cc30b1f12 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.267804] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Releasing lock "refresh_cache-dab6d2c5-8c04-4977-a409-b384037eabce" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.268406] env[61987]: DEBUG nova.objects.instance [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lazy-loading 'flavor' on Instance uuid dab6d2c5-8c04-4977-a409-b384037eabce {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.269823] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062241, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.271243] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1031.271243] env[61987]: value = "task-1062242" [ 1031.271243] env[61987]: _type = "Task" [ 1031.271243] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.284169] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c2e8e0-de37-a43b-43ab-67deb6783249, 'name': SearchDatastore_Task, 'duration_secs': 0.008483} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.287983] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.288301] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] bf1123bf-80e6-4be3-804f-e13b906ee44d/bf1123bf-80e6-4be3-804f-e13b906ee44d.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1031.288850] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062242, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.289092] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0798af47-2b3e-4a2d-8cc5-cff1d262afb7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.294870] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 1031.294870] env[61987]: value = "task-1062243" [ 1031.294870] env[61987]: _type = "Task" [ 1031.294870] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.303190] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062243, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.313218] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f621c9f0-f567-456e-a24b-f1a8d3ab4846 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "0cba5e51-7e0a-409b-a7b2-612ab8cf5689" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.510s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.644972] env[61987]: INFO nova.compute.manager [-] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Took 1.04 seconds to deallocate network for instance. [ 1031.677896] env[61987]: DEBUG oslo_concurrency.lockutils [None req-71ba2865-5169-40f4-8456-f2f4d5d69ecd tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "b172195a-79d2-4f24-b9df-320a07de965e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.056s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.781021] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062241, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.786627] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0329933-e988-4875-9044-9925d9ef1065 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.795587] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062242, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069428} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.813877] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1031.814367] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1031.818439] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e85e45f0-4f52-4c9d-a5ab-9cb73bb0eb88 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.821812] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-476e6454-71ae-4c65-87fd-ec4d6c262d10 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.851334] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] f3773cda-41e8-4804-9319-c7e458112ddf/f3773cda-41e8-4804-9319-c7e458112ddf.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1031.858508] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8d27168f-64ae-4155-a45a-e1b6109e84e2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.876612] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062243, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.877222] env[61987]: DEBUG oslo_vmware.api [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1031.877222] env[61987]: value = "task-1062244" [ 1031.877222] env[61987]: _type = "Task" [ 1031.877222] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.884456] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1031.884456] env[61987]: value = "task-1062245" [ 1031.884456] env[61987]: _type = "Task" [ 1031.884456] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.887795] env[61987]: DEBUG oslo_vmware.api [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062244, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.895564] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062245, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.153876] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.154024] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.154215] env[61987]: DEBUG nova.objects.instance [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Lazy-loading 'resources' on Instance uuid 0da22c35-f11d-4498-bdb4-31bfd4c364e6 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.224106] env[61987]: DEBUG nova.network.neutron [req-cec50c54-bebe-44e6-bdfc-b23c21c0427e req-9982a308-0d1c-4f73-a271-c4bf6829c37d service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Updated VIF entry in instance network info cache for port d96dede6-c439-4a14-a4d1-62c44e03ea11. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1032.224106] env[61987]: DEBUG nova.network.neutron [req-cec50c54-bebe-44e6-bdfc-b23c21c0427e req-9982a308-0d1c-4f73-a271-c4bf6829c37d service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Updating instance_info_cache with network_info: [{"id": "d96dede6-c439-4a14-a4d1-62c44e03ea11", "address": "fa:16:3e:d2:e0:b1", "network": {"id": "4ed4fbdc-05b0-44f8-a3d5-b7b288abd5e0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2067313775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a234a62da0e498fbe613fbcaaec3201", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd96dede6-c4", "ovs_interfaceid": "d96dede6-c439-4a14-a4d1-62c44e03ea11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.273157] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062241, 'name': CreateVM_Task, 'duration_secs': 0.559251} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.273350] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1032.274093] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.274272] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.274608] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1032.274863] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a288a65-793c-4d27-94b7-003e001fd3a2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.279578] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1032.279578] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]527ce0fb-5f6b-99bc-6218-997075eea5be" [ 1032.279578] env[61987]: _type = "Task" [ 1032.279578] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.287465] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]527ce0fb-5f6b-99bc-6218-997075eea5be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.324846] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062243, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.606994} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.324846] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] bf1123bf-80e6-4be3-804f-e13b906ee44d/bf1123bf-80e6-4be3-804f-e13b906ee44d.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1032.324846] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1032.324846] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ede3da14-c584-4284-ad7d-31e27f5709be {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.330193] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 1032.330193] env[61987]: value = "task-1062246" [ 1032.330193] env[61987]: _type = "Task" [ 1032.330193] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.338844] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062246, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.387793] env[61987]: DEBUG oslo_vmware.api [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062244, 'name': PowerOffVM_Task, 'duration_secs': 0.220305} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.387793] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1032.391904] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Reconfiguring VM instance instance-00000069 to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1032.392308] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-527a633f-a3ce-4772-891e-629527efbb41 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.414119] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062245, 'name': ReconfigVM_Task, 'duration_secs': 0.319911} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.415360] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Reconfigured VM instance instance-0000006a to attach disk [datastore1] f3773cda-41e8-4804-9319-c7e458112ddf/f3773cda-41e8-4804-9319-c7e458112ddf.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1032.416114] env[61987]: DEBUG oslo_vmware.api [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1032.416114] env[61987]: value = "task-1062247" [ 1032.416114] env[61987]: _type = "Task" [ 1032.416114] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.416361] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-95fd2a0e-c630-4c8c-ac9e-0710918630f0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.425745] env[61987]: DEBUG oslo_vmware.api [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062247, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.427019] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1032.427019] env[61987]: value = "task-1062248" [ 1032.427019] env[61987]: _type = "Task" [ 1032.427019] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.434782] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062248, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.635101] env[61987]: DEBUG nova.compute.manager [req-8fd6db78-e1dc-4a2e-b53f-c57a44de6a31 req-bdd1998f-68df-462c-a5e8-f1e8f21aea50 service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Received event network-changed-fe549122-49ad-4628-b106-216d787be2d4 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1032.636101] env[61987]: DEBUG nova.compute.manager [req-8fd6db78-e1dc-4a2e-b53f-c57a44de6a31 req-bdd1998f-68df-462c-a5e8-f1e8f21aea50 service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Refreshing instance network info cache due to event network-changed-fe549122-49ad-4628-b106-216d787be2d4. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1032.636370] env[61987]: DEBUG oslo_concurrency.lockutils [req-8fd6db78-e1dc-4a2e-b53f-c57a44de6a31 req-bdd1998f-68df-462c-a5e8-f1e8f21aea50 service nova] Acquiring lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.636544] env[61987]: DEBUG oslo_concurrency.lockutils [req-8fd6db78-e1dc-4a2e-b53f-c57a44de6a31 req-bdd1998f-68df-462c-a5e8-f1e8f21aea50 service nova] Acquired lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.636721] env[61987]: DEBUG nova.network.neutron [req-8fd6db78-e1dc-4a2e-b53f-c57a44de6a31 req-bdd1998f-68df-462c-a5e8-f1e8f21aea50 service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Refreshing network info cache for port fe549122-49ad-4628-b106-216d787be2d4 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1032.704909] env[61987]: INFO nova.compute.manager [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Rebuilding instance [ 1032.728412] env[61987]: DEBUG oslo_concurrency.lockutils [req-cec50c54-bebe-44e6-bdfc-b23c21c0427e req-9982a308-0d1c-4f73-a271-c4bf6829c37d service nova] Releasing lock "refresh_cache-7002962c-76ed-4a8d-be2a-b2b118847fea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.756905] env[61987]: DEBUG nova.compute.manager [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1032.757887] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f426d6e-abe6-4698-9bb6-5e7f11d0b4ae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.794088] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]527ce0fb-5f6b-99bc-6218-997075eea5be, 'name': SearchDatastore_Task, 'duration_secs': 0.008937} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.794578] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.794771] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1032.794963] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.795130] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.795343] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1032.795623] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-884633af-c4d8-4176-bdb3-205b92ed4792 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.808052] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1032.809121] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1032.809121] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcf9191c-5495-4162-b03e-94f3b2e21c75 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.813885] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1032.813885] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5285698b-3db2-6c5d-a850-ee6c557ad8fa" [ 1032.813885] env[61987]: _type = "Task" [ 1032.813885] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.823643] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5285698b-3db2-6c5d-a850-ee6c557ad8fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.845162] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062246, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084224} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.845280] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1032.846054] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15b976c7-c62b-4585-a776-26fca083922c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.871776] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] bf1123bf-80e6-4be3-804f-e13b906ee44d/bf1123bf-80e6-4be3-804f-e13b906ee44d.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1032.874864] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bed3722a-06bd-4ba6-b637-01f5a45dffad {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.895901] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 1032.895901] env[61987]: value = "task-1062249" [ 1032.895901] env[61987]: _type = "Task" [ 1032.895901] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.904645] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062249, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.927620] env[61987]: DEBUG oslo_vmware.api [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062247, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.932730] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c57d36fc-a8d7-4fe3-a835-76d8b86437a7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.938262] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062248, 'name': Rename_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.942506] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58779934-72db-4fdb-9bbe-eded310fbdaf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.972228] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e683c8c-df60-4c5b-8f4e-d3aaa1c464e8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.980322] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b18a333-577e-441e-b464-aa4c8c76dc58 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.995729] env[61987]: DEBUG nova.compute.provider_tree [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1033.325251] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5285698b-3db2-6c5d-a850-ee6c557ad8fa, 'name': SearchDatastore_Task, 'duration_secs': 0.00965} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.326153] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ff908c4-da67-40d5-94a8-2a58f59de3f9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.331582] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1033.331582] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52650b97-2491-001e-c427-6a01657e1ad2" [ 1033.331582] env[61987]: _type = "Task" [ 1033.331582] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.339611] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52650b97-2491-001e-c427-6a01657e1ad2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.361988] env[61987]: DEBUG nova.network.neutron [req-8fd6db78-e1dc-4a2e-b53f-c57a44de6a31 req-bdd1998f-68df-462c-a5e8-f1e8f21aea50 service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Updated VIF entry in instance network info cache for port fe549122-49ad-4628-b106-216d787be2d4. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1033.362391] env[61987]: DEBUG nova.network.neutron [req-8fd6db78-e1dc-4a2e-b53f-c57a44de6a31 req-bdd1998f-68df-462c-a5e8-f1e8f21aea50 service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Updating instance_info_cache with network_info: [{"id": "fe549122-49ad-4628-b106-216d787be2d4", "address": "fa:16:3e:76:0c:4d", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe549122-49", "ovs_interfaceid": "fe549122-49ad-4628-b106-216d787be2d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.405918] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062249, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.429893] env[61987]: DEBUG oslo_vmware.api [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062247, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.440010] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062248, 'name': Rename_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.498641] env[61987]: DEBUG nova.scheduler.client.report [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1033.775380] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1033.775695] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a04e5e5d-17fe-4dca-bdea-d79c199aaa78 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.782507] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1033.782507] env[61987]: value = "task-1062250" [ 1033.782507] env[61987]: _type = "Task" [ 1033.782507] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.790295] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062250, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.816631] env[61987]: DEBUG nova.compute.manager [req-048f4089-e6bb-4116-81ed-671761b0acfa req-8a405569-bccb-403f-85aa-1cc09b505594 service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Received event network-changed-b739a840-515d-46e6-b889-5e9f8ec9ad71 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1033.816907] env[61987]: DEBUG nova.compute.manager [req-048f4089-e6bb-4116-81ed-671761b0acfa req-8a405569-bccb-403f-85aa-1cc09b505594 service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Refreshing instance network info cache due to event network-changed-b739a840-515d-46e6-b889-5e9f8ec9ad71. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1033.817106] env[61987]: DEBUG oslo_concurrency.lockutils [req-048f4089-e6bb-4116-81ed-671761b0acfa req-8a405569-bccb-403f-85aa-1cc09b505594 service nova] Acquiring lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.817345] env[61987]: DEBUG oslo_concurrency.lockutils [req-048f4089-e6bb-4116-81ed-671761b0acfa req-8a405569-bccb-403f-85aa-1cc09b505594 service nova] Acquired lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.817449] env[61987]: DEBUG nova.network.neutron [req-048f4089-e6bb-4116-81ed-671761b0acfa req-8a405569-bccb-403f-85aa-1cc09b505594 service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Refreshing network info cache for port b739a840-515d-46e6-b889-5e9f8ec9ad71 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1033.842115] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52650b97-2491-001e-c427-6a01657e1ad2, 'name': SearchDatastore_Task, 'duration_secs': 0.009564} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.842424] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.842682] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 7002962c-76ed-4a8d-be2a-b2b118847fea/7002962c-76ed-4a8d-be2a-b2b118847fea.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1033.842950] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c4a3ad3e-f0a6-4760-93a1-f143a8376a23 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.850199] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1033.850199] env[61987]: value = "task-1062251" [ 1033.850199] env[61987]: _type = "Task" [ 1033.850199] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.858465] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062251, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.865489] env[61987]: DEBUG oslo_concurrency.lockutils [req-8fd6db78-e1dc-4a2e-b53f-c57a44de6a31 req-bdd1998f-68df-462c-a5e8-f1e8f21aea50 service nova] Releasing lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.865667] env[61987]: DEBUG nova.compute.manager [req-8fd6db78-e1dc-4a2e-b53f-c57a44de6a31 req-bdd1998f-68df-462c-a5e8-f1e8f21aea50 service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Received event network-changed-b739a840-515d-46e6-b889-5e9f8ec9ad71 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1033.865852] env[61987]: DEBUG nova.compute.manager [req-8fd6db78-e1dc-4a2e-b53f-c57a44de6a31 req-bdd1998f-68df-462c-a5e8-f1e8f21aea50 service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Refreshing instance network info cache due to event network-changed-b739a840-515d-46e6-b889-5e9f8ec9ad71. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1033.866065] env[61987]: DEBUG oslo_concurrency.lockutils [req-8fd6db78-e1dc-4a2e-b53f-c57a44de6a31 req-bdd1998f-68df-462c-a5e8-f1e8f21aea50 service nova] Acquiring lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.905956] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062249, 'name': ReconfigVM_Task, 'duration_secs': 0.755666} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.906227] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Reconfigured VM instance instance-0000006d to attach disk [datastore1] bf1123bf-80e6-4be3-804f-e13b906ee44d/bf1123bf-80e6-4be3-804f-e13b906ee44d.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1033.906873] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b6c449f4-df73-4eae-862c-ca5909eeab25 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.912712] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 1033.912712] env[61987]: value = "task-1062252" [ 1033.912712] env[61987]: _type = "Task" [ 1033.912712] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.922170] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062252, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.930462] env[61987]: DEBUG oslo_vmware.api [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062247, 'name': ReconfigVM_Task, 'duration_secs': 1.261137} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.668274] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Reconfigured VM instance instance-00000069 to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1034.668494] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1034.669770] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.516s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.673648] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-95a1c343-e3b1-470a-b245-5f7e1337976f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.688384] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062248, 'name': Rename_Task, 'duration_secs': 1.035269} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.688548] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1034.688791] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9d7e393e-23e8-42d9-9e8b-ef2bdb513261 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.691491] env[61987]: DEBUG oslo_vmware.api [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1034.691491] env[61987]: value = "task-1062253" [ 1034.691491] env[61987]: _type = "Task" [ 1034.691491] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.701225] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062252, 'name': Rename_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.701489] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062251, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.508652} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.702492] env[61987]: INFO nova.scheduler.client.report [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Deleted allocations for instance 0da22c35-f11d-4498-bdb4-31bfd4c364e6 [ 1034.706475] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 7002962c-76ed-4a8d-be2a-b2b118847fea/7002962c-76ed-4a8d-be2a-b2b118847fea.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1034.707069] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1034.707069] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] VM already powered off {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1034.707300] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1034.711400] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5984b33c-82c5-4879-a150-06bbb476a0f9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.714055] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9eb25a4-2508-4391-b1b5-68c54d20a255 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.716928] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1034.716928] env[61987]: value = "task-1062254" [ 1034.716928] env[61987]: _type = "Task" [ 1034.716928] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.729576] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1034.730245] env[61987]: DEBUG oslo_vmware.api [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062253, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.731775] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2fb95d6c-94cf-4664-8245-c25da76f5785 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.734393] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1034.734393] env[61987]: value = "task-1062255" [ 1034.734393] env[61987]: _type = "Task" [ 1034.734393] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.739468] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062254, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.750130] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062255, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.811572] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1034.811572] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1034.811868] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Deleting the datastore file [datastore2] b172195a-79d2-4f24-b9df-320a07de965e {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1034.814444] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-63a543f6-ae26-4fd0-867d-12f4dbf73e56 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.820568] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1034.820568] env[61987]: value = "task-1062257" [ 1034.820568] env[61987]: _type = "Task" [ 1034.820568] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.828577] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062257, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.935117] env[61987]: DEBUG nova.network.neutron [req-048f4089-e6bb-4116-81ed-671761b0acfa req-8a405569-bccb-403f-85aa-1cc09b505594 service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Updated VIF entry in instance network info cache for port b739a840-515d-46e6-b889-5e9f8ec9ad71. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1034.935565] env[61987]: DEBUG nova.network.neutron [req-048f4089-e6bb-4116-81ed-671761b0acfa req-8a405569-bccb-403f-85aa-1cc09b505594 service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Updating instance_info_cache with network_info: [{"id": "b739a840-515d-46e6-b889-5e9f8ec9ad71", "address": "fa:16:3e:cb:9f:9b", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb739a840-51", "ovs_interfaceid": "b739a840-515d-46e6-b889-5e9f8ec9ad71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.182802] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062252, 'name': Rename_Task, 'duration_secs': 0.893917} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.182802] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1035.183104] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aedc6528-b09a-429d-ae17-b6e1c6642f22 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.189127] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 1035.189127] env[61987]: value = "task-1062258" [ 1035.189127] env[61987]: _type = "Task" [ 1035.189127] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.196772] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062258, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.210357] env[61987]: DEBUG oslo_vmware.api [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062253, 'name': PowerOnVM_Task, 'duration_secs': 0.453479} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.210626] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1035.210866] env[61987]: DEBUG nova.compute.manager [None req-b35b95f4-6088-479e-b782-c059bf69a99a tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1035.211679] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43473653-da40-4e50-a356-02a5365e8153 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.226412] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0abc2a2e-114e-4d82-8cb8-14f515d7fe1a tempest-ServerShowV254Test-1364450372 tempest-ServerShowV254Test-1364450372-project-member] Lock "0da22c35-f11d-4498-bdb4-31bfd4c364e6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.392s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.234940] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062254, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.248093] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062255, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073369} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.248386] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1035.249167] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8afd75f-90d0-4f9c-846e-5bd96d2753e7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.272236] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] 7002962c-76ed-4a8d-be2a-b2b118847fea/7002962c-76ed-4a8d-be2a-b2b118847fea.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1035.272863] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16823650-ded5-4831-968d-9db6032acfb8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.293309] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1035.293309] env[61987]: value = "task-1062259" [ 1035.293309] env[61987]: _type = "Task" [ 1035.293309] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.303259] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062259, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.332121] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062257, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.170078} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.332396] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1035.332596] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1035.332736] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1035.439236] env[61987]: DEBUG oslo_concurrency.lockutils [req-048f4089-e6bb-4116-81ed-671761b0acfa req-8a405569-bccb-403f-85aa-1cc09b505594 service nova] Releasing lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.439748] env[61987]: DEBUG oslo_concurrency.lockutils [req-8fd6db78-e1dc-4a2e-b53f-c57a44de6a31 req-bdd1998f-68df-462c-a5e8-f1e8f21aea50 service nova] Acquired lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.439972] env[61987]: DEBUG nova.network.neutron [req-8fd6db78-e1dc-4a2e-b53f-c57a44de6a31 req-bdd1998f-68df-462c-a5e8-f1e8f21aea50 service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Refreshing network info cache for port b739a840-515d-46e6-b889-5e9f8ec9ad71 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1035.702635] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062258, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.734668] env[61987]: DEBUG oslo_vmware.api [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062254, 'name': PowerOnVM_Task, 'duration_secs': 0.635165} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.734942] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1035.735172] env[61987]: DEBUG nova.compute.manager [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1035.735952] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-707a128a-1dab-4e60-aa35-780a0a7bd5e7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.803267] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062259, 'name': ReconfigVM_Task, 'duration_secs': 0.318996} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.803591] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Reconfigured VM instance instance-0000006e to attach disk [datastore1] 7002962c-76ed-4a8d-be2a-b2b118847fea/7002962c-76ed-4a8d-be2a-b2b118847fea.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1035.804215] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9af7aff6-1313-430a-bf81-c8700a110e40 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.809711] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1035.809711] env[61987]: value = "task-1062260" [ 1035.809711] env[61987]: _type = "Task" [ 1035.809711] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.817187] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062260, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.844780] env[61987]: DEBUG nova.compute.manager [req-cf595430-d0e6-41d4-8878-2f3fce9845c3 req-1c9d3d28-d3a7-4297-a54d-6ada95e9e561 service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Received event network-changed-fe549122-49ad-4628-b106-216d787be2d4 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1035.845065] env[61987]: DEBUG nova.compute.manager [req-cf595430-d0e6-41d4-8878-2f3fce9845c3 req-1c9d3d28-d3a7-4297-a54d-6ada95e9e561 service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Refreshing instance network info cache due to event network-changed-fe549122-49ad-4628-b106-216d787be2d4. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1035.845565] env[61987]: DEBUG oslo_concurrency.lockutils [req-cf595430-d0e6-41d4-8878-2f3fce9845c3 req-1c9d3d28-d3a7-4297-a54d-6ada95e9e561 service nova] Acquiring lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.845565] env[61987]: DEBUG oslo_concurrency.lockutils [req-cf595430-d0e6-41d4-8878-2f3fce9845c3 req-1c9d3d28-d3a7-4297-a54d-6ada95e9e561 service nova] Acquired lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.845762] env[61987]: DEBUG nova.network.neutron [req-cf595430-d0e6-41d4-8878-2f3fce9845c3 req-1c9d3d28-d3a7-4297-a54d-6ada95e9e561 service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Refreshing network info cache for port fe549122-49ad-4628-b106-216d787be2d4 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1036.200492] env[61987]: DEBUG oslo_vmware.api [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062258, 'name': PowerOnVM_Task, 'duration_secs': 1.004526} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.201383] env[61987]: DEBUG nova.network.neutron [req-8fd6db78-e1dc-4a2e-b53f-c57a44de6a31 req-bdd1998f-68df-462c-a5e8-f1e8f21aea50 service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Updated VIF entry in instance network info cache for port b739a840-515d-46e6-b889-5e9f8ec9ad71. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1036.201721] env[61987]: DEBUG nova.network.neutron [req-8fd6db78-e1dc-4a2e-b53f-c57a44de6a31 req-bdd1998f-68df-462c-a5e8-f1e8f21aea50 service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Updating instance_info_cache with network_info: [{"id": "b739a840-515d-46e6-b889-5e9f8ec9ad71", "address": "fa:16:3e:cb:9f:9b", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb739a840-51", "ovs_interfaceid": "b739a840-515d-46e6-b889-5e9f8ec9ad71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.203019] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1036.203147] env[61987]: INFO nova.compute.manager [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Took 10.03 seconds to spawn the instance on the hypervisor. [ 1036.203696] env[61987]: DEBUG nova.compute.manager [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1036.205094] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f00cfbf8-18bd-4ff2-b778-4ca216d7233b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.254263] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.254379] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.254469] env[61987]: DEBUG nova.objects.instance [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61987) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1036.320427] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062260, 'name': Rename_Task, 'duration_secs': 0.147488} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.320723] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1036.321062] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b1f872a8-acbf-4734-a663-094409baf45b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.327731] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1036.327731] env[61987]: value = "task-1062261" [ 1036.327731] env[61987]: _type = "Task" [ 1036.327731] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.336526] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062261, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.378762] env[61987]: DEBUG nova.virt.hardware [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1036.379054] env[61987]: DEBUG nova.virt.hardware [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1036.379236] env[61987]: DEBUG nova.virt.hardware [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1036.379432] env[61987]: DEBUG nova.virt.hardware [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1036.379586] env[61987]: DEBUG nova.virt.hardware [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1036.379762] env[61987]: DEBUG nova.virt.hardware [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1036.379981] env[61987]: DEBUG nova.virt.hardware [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1036.380958] env[61987]: DEBUG nova.virt.hardware [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1036.380958] env[61987]: DEBUG nova.virt.hardware [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1036.380958] env[61987]: DEBUG nova.virt.hardware [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1036.380958] env[61987]: DEBUG nova.virt.hardware [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1036.381590] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2febee2-3730-4437-af14-ffa92aa99c3f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.391941] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac915fa8-62bf-4ef7-9b42-4f250999bbb8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.405073] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:20:28:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'be8bd197-4b2b-46e7-88ea-2554b0438584', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2a9d7219-a935-4e81-9278-7abce87993fe', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1036.412584] env[61987]: DEBUG oslo.service.loopingcall [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.412828] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1036.413057] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-033e2afd-f83d-4c82-a0af-1ef333afe5ac {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.435888] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1036.435888] env[61987]: value = "task-1062262" [ 1036.435888] env[61987]: _type = "Task" [ 1036.435888] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.444234] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062262, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.690045] env[61987]: DEBUG nova.network.neutron [req-cf595430-d0e6-41d4-8878-2f3fce9845c3 req-1c9d3d28-d3a7-4297-a54d-6ada95e9e561 service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Updated VIF entry in instance network info cache for port fe549122-49ad-4628-b106-216d787be2d4. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1036.690501] env[61987]: DEBUG nova.network.neutron [req-cf595430-d0e6-41d4-8878-2f3fce9845c3 req-1c9d3d28-d3a7-4297-a54d-6ada95e9e561 service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Updating instance_info_cache with network_info: [{"id": "fe549122-49ad-4628-b106-216d787be2d4", "address": "fa:16:3e:76:0c:4d", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe549122-49", "ovs_interfaceid": "fe549122-49ad-4628-b106-216d787be2d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.705046] env[61987]: DEBUG oslo_concurrency.lockutils [req-8fd6db78-e1dc-4a2e-b53f-c57a44de6a31 req-bdd1998f-68df-462c-a5e8-f1e8f21aea50 service nova] Releasing lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.721763] env[61987]: INFO nova.compute.manager [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Took 20.87 seconds to build instance. [ 1036.842166] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062261, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.932580] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquiring lock "dab6d2c5-8c04-4977-a409-b384037eabce" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.932933] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lock "dab6d2c5-8c04-4977-a409-b384037eabce" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.933117] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquiring lock "dab6d2c5-8c04-4977-a409-b384037eabce-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.933302] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lock "dab6d2c5-8c04-4977-a409-b384037eabce-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.933483] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lock "dab6d2c5-8c04-4977-a409-b384037eabce-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.936316] env[61987]: INFO nova.compute.manager [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Terminating instance [ 1036.947799] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062262, 'name': CreateVM_Task, 'duration_secs': 0.407423} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.948041] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1036.951863] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.951863] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.951863] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1036.951863] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f2db09c-8145-439a-b3d7-b7842d447dd5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.959177] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1036.959177] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52fce2ad-3924-ff70-6bc5-22be434e1ff6" [ 1036.959177] env[61987]: _type = "Task" [ 1036.959177] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.972937] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52fce2ad-3924-ff70-6bc5-22be434e1ff6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.193815] env[61987]: DEBUG oslo_concurrency.lockutils [req-cf595430-d0e6-41d4-8878-2f3fce9845c3 req-1c9d3d28-d3a7-4297-a54d-6ada95e9e561 service nova] Releasing lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.224529] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51985d87-51f1-4472-ba4f-4c932ecb535d tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "bf1123bf-80e6-4be3-804f-e13b906ee44d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.381s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.265084] env[61987]: DEBUG oslo_concurrency.lockutils [None req-7f14ea94-4d0e-4097-9aec-d463105bdc7d tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.338821] env[61987]: DEBUG oslo_vmware.api [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062261, 'name': PowerOnVM_Task, 'duration_secs': 0.609096} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.339138] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1037.339349] env[61987]: INFO nova.compute.manager [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Took 8.79 seconds to spawn the instance on the hypervisor. [ 1037.339530] env[61987]: DEBUG nova.compute.manager [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1037.340327] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ef4d1fa-826e-4c50-81e4-4bd21c7c132b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.446229] env[61987]: DEBUG nova.compute.manager [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1037.446525] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1037.447415] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b45c018c-f3d6-41dc-af7a-58b041509782 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.460876] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1037.464778] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3c778ada-4736-4fc9-bce5-29489b5960d8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.476369] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52fce2ad-3924-ff70-6bc5-22be434e1ff6, 'name': SearchDatastore_Task, 'duration_secs': 0.034287} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.477935] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.478210] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1037.478461] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.478617] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.478801] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1037.479143] env[61987]: DEBUG oslo_vmware.api [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1037.479143] env[61987]: value = "task-1062263" [ 1037.479143] env[61987]: _type = "Task" [ 1037.479143] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.479415] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-01cd4d5f-9615-4dce-9324-43589f22246a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.494017] env[61987]: DEBUG oslo_vmware.api [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062263, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.495295] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1037.495472] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1037.496194] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b1b93e3-a06b-48ae-b332-926f574b8ecf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.501534] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1037.501534] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52c81525-6b02-6097-b196-06ef5787a7a1" [ 1037.501534] env[61987]: _type = "Task" [ 1037.501534] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.509832] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c81525-6b02-6097-b196-06ef5787a7a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.526778] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "f3773cda-41e8-4804-9319-c7e458112ddf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.527155] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "f3773cda-41e8-4804-9319-c7e458112ddf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.527268] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "f3773cda-41e8-4804-9319-c7e458112ddf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.527453] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "f3773cda-41e8-4804-9319-c7e458112ddf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.527629] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "f3773cda-41e8-4804-9319-c7e458112ddf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.530099] env[61987]: INFO nova.compute.manager [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Terminating instance [ 1037.818729] env[61987]: DEBUG nova.compute.manager [req-e83e1a58-a78b-4547-85ce-d5e91854ff7b req-94b62f9e-abf8-4611-a2d2-2587e06a8039 service nova] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Received event network-changed-a46b1587-a8f9-4552-8d05-57882397c3f5 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1037.819297] env[61987]: DEBUG nova.compute.manager [req-e83e1a58-a78b-4547-85ce-d5e91854ff7b req-94b62f9e-abf8-4611-a2d2-2587e06a8039 service nova] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Refreshing instance network info cache due to event network-changed-a46b1587-a8f9-4552-8d05-57882397c3f5. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1037.819359] env[61987]: DEBUG oslo_concurrency.lockutils [req-e83e1a58-a78b-4547-85ce-d5e91854ff7b req-94b62f9e-abf8-4611-a2d2-2587e06a8039 service nova] Acquiring lock "refresh_cache-bf1123bf-80e6-4be3-804f-e13b906ee44d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.819548] env[61987]: DEBUG oslo_concurrency.lockutils [req-e83e1a58-a78b-4547-85ce-d5e91854ff7b req-94b62f9e-abf8-4611-a2d2-2587e06a8039 service nova] Acquired lock "refresh_cache-bf1123bf-80e6-4be3-804f-e13b906ee44d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.819783] env[61987]: DEBUG nova.network.neutron [req-e83e1a58-a78b-4547-85ce-d5e91854ff7b req-94b62f9e-abf8-4611-a2d2-2587e06a8039 service nova] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Refreshing network info cache for port a46b1587-a8f9-4552-8d05-57882397c3f5 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1037.860802] env[61987]: INFO nova.compute.manager [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Took 16.19 seconds to build instance. [ 1038.005385] env[61987]: DEBUG oslo_vmware.api [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062263, 'name': PowerOffVM_Task, 'duration_secs': 0.265143} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.010072] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1038.010398] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1038.010626] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fd5e38a6-a723-46e2-a68b-fdd9fc261f8e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.021772] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c81525-6b02-6097-b196-06ef5787a7a1, 'name': SearchDatastore_Task, 'duration_secs': 0.011793} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.021772] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddead3ea-be91-4f9c-9162-eaab8aed62dd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.028385] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1038.028385] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52a63d9a-4be2-fdce-a334-5114175d2973" [ 1038.028385] env[61987]: _type = "Task" [ 1038.028385] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.034132] env[61987]: DEBUG nova.compute.manager [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1038.034346] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1038.039255] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74a99fa8-7026-49c5-b079-b33d44eb2148 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.042196] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52a63d9a-4be2-fdce-a334-5114175d2973, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.046802] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1038.047259] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-787b43e1-99ec-473f-a777-4b8ac5f71387 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.053560] env[61987]: DEBUG oslo_vmware.api [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1038.053560] env[61987]: value = "task-1062265" [ 1038.053560] env[61987]: _type = "Task" [ 1038.053560] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.061744] env[61987]: DEBUG oslo_vmware.api [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062265, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.082426] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1038.082676] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1038.082864] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Deleting the datastore file [datastore1] dab6d2c5-8c04-4977-a409-b384037eabce {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1038.083165] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e2446844-9bfd-477b-8e3c-44aba9a59c25 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.090651] env[61987]: DEBUG oslo_vmware.api [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1038.090651] env[61987]: value = "task-1062266" [ 1038.090651] env[61987]: _type = "Task" [ 1038.090651] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.099726] env[61987]: DEBUG oslo_vmware.api [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062266, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.364944] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c76926ce-31e0-41fe-8815-9481c730d665 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "7002962c-76ed-4a8d-be2a-b2b118847fea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.708s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.539938] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52a63d9a-4be2-fdce-a334-5114175d2973, 'name': SearchDatastore_Task, 'duration_secs': 0.018748} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.540434] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.541695] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] b172195a-79d2-4f24-b9df-320a07de965e/b172195a-79d2-4f24-b9df-320a07de965e.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1038.542177] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cb9c4b66-c93b-4d0f-abb9-f3e93fd73624 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.554118] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1038.554118] env[61987]: value = "task-1062267" [ 1038.554118] env[61987]: _type = "Task" [ 1038.554118] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.567285] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062267, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.571590] env[61987]: DEBUG oslo_vmware.api [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062265, 'name': PowerOffVM_Task, 'duration_secs': 0.169184} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.571590] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1038.571843] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1038.572208] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7b755dd8-d393-4881-af07-3f5ebcf90f05 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.603080] env[61987]: DEBUG oslo_vmware.api [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062266, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.3732} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.604029] env[61987]: DEBUG nova.network.neutron [req-e83e1a58-a78b-4547-85ce-d5e91854ff7b req-94b62f9e-abf8-4611-a2d2-2587e06a8039 service nova] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Updated VIF entry in instance network info cache for port a46b1587-a8f9-4552-8d05-57882397c3f5. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1038.604368] env[61987]: DEBUG nova.network.neutron [req-e83e1a58-a78b-4547-85ce-d5e91854ff7b req-94b62f9e-abf8-4611-a2d2-2587e06a8039 service nova] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Updating instance_info_cache with network_info: [{"id": "a46b1587-a8f9-4552-8d05-57882397c3f5", "address": "fa:16:3e:b2:84:05", "network": {"id": "26d453d1-8edf-45e1-9e49-2a3225de6633", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1400564876-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b77fbbfc7f6437083f9f30e8fdeeb35", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3734b156-0f7d-4721-b23c-d000412ec2eb", "external-id": "nsx-vlan-transportzone-560", "segmentation_id": 560, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa46b1587-a8", "ovs_interfaceid": "a46b1587-a8f9-4552-8d05-57882397c3f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.605616] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1038.605785] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1038.605964] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1038.606283] env[61987]: INFO nova.compute.manager [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1038.606549] env[61987]: DEBUG oslo.service.loopingcall [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1038.606761] env[61987]: DEBUG nova.compute.manager [-] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1038.606855] env[61987]: DEBUG nova.network.neutron [-] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1038.634135] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1038.634420] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1038.634622] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Deleting the datastore file [datastore1] f3773cda-41e8-4804-9319-c7e458112ddf {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1038.635330] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e223958b-b9c0-4441-a80b-d7767e107b63 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.642751] env[61987]: DEBUG oslo_vmware.api [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1038.642751] env[61987]: value = "task-1062269" [ 1038.642751] env[61987]: _type = "Task" [ 1038.642751] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.652755] env[61987]: DEBUG oslo_vmware.api [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062269, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.077749] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062267, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.110483] env[61987]: DEBUG oslo_concurrency.lockutils [req-e83e1a58-a78b-4547-85ce-d5e91854ff7b req-94b62f9e-abf8-4611-a2d2-2587e06a8039 service nova] Releasing lock "refresh_cache-bf1123bf-80e6-4be3-804f-e13b906ee44d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.158282] env[61987]: DEBUG oslo_vmware.api [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062269, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.210164} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.158817] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1039.158923] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1039.159170] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1039.159426] env[61987]: INFO nova.compute.manager [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1039.159761] env[61987]: DEBUG oslo.service.loopingcall [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.160173] env[61987]: DEBUG nova.compute.manager [-] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1039.160173] env[61987]: DEBUG nova.network.neutron [-] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1039.298831] env[61987]: DEBUG nova.compute.manager [req-95731062-4ce5-41e5-a9d4-997767d44e1d req-6b300f60-adf1-44ee-bd96-efe597759e6e service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Received event network-changed-d96dede6-c439-4a14-a4d1-62c44e03ea11 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1039.299055] env[61987]: DEBUG nova.compute.manager [req-95731062-4ce5-41e5-a9d4-997767d44e1d req-6b300f60-adf1-44ee-bd96-efe597759e6e service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Refreshing instance network info cache due to event network-changed-d96dede6-c439-4a14-a4d1-62c44e03ea11. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1039.299288] env[61987]: DEBUG oslo_concurrency.lockutils [req-95731062-4ce5-41e5-a9d4-997767d44e1d req-6b300f60-adf1-44ee-bd96-efe597759e6e service nova] Acquiring lock "refresh_cache-7002962c-76ed-4a8d-be2a-b2b118847fea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.299476] env[61987]: DEBUG oslo_concurrency.lockutils [req-95731062-4ce5-41e5-a9d4-997767d44e1d req-6b300f60-adf1-44ee-bd96-efe597759e6e service nova] Acquired lock "refresh_cache-7002962c-76ed-4a8d-be2a-b2b118847fea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.299592] env[61987]: DEBUG nova.network.neutron [req-95731062-4ce5-41e5-a9d4-997767d44e1d req-6b300f60-adf1-44ee-bd96-efe597759e6e service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Refreshing network info cache for port d96dede6-c439-4a14-a4d1-62c44e03ea11 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1039.567940] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062267, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.660451} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.568249] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] b172195a-79d2-4f24-b9df-320a07de965e/b172195a-79d2-4f24-b9df-320a07de965e.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1039.568512] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1039.569101] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3b9fb2bb-5ec7-49fe-b6f4-a94c17fdee40 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.576862] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1039.576862] env[61987]: value = "task-1062270" [ 1039.576862] env[61987]: _type = "Task" [ 1039.576862] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.585014] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062270, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.587862] env[61987]: DEBUG nova.network.neutron [-] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.857411] env[61987]: DEBUG nova.compute.manager [req-5bb2d1d0-cfcf-43b3-80c6-14207b954ca6 req-4f06c3dc-40ae-48f3-94f1-2ee931f8e005 service nova] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Received event network-vif-deleted-05cdb86d-67c4-4b34-9cd3-4e4b3ed04d58 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1039.857649] env[61987]: DEBUG nova.compute.manager [req-5bb2d1d0-cfcf-43b3-80c6-14207b954ca6 req-4f06c3dc-40ae-48f3-94f1-2ee931f8e005 service nova] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Received event network-vif-deleted-b5999bb0-480a-45bd-b07c-fbb91985421e {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1039.857834] env[61987]: INFO nova.compute.manager [req-5bb2d1d0-cfcf-43b3-80c6-14207b954ca6 req-4f06c3dc-40ae-48f3-94f1-2ee931f8e005 service nova] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Neutron deleted interface b5999bb0-480a-45bd-b07c-fbb91985421e; detaching it from the instance and deleting it from the info cache [ 1039.858020] env[61987]: DEBUG nova.network.neutron [req-5bb2d1d0-cfcf-43b3-80c6-14207b954ca6 req-4f06c3dc-40ae-48f3-94f1-2ee931f8e005 service nova] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.086527] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062270, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073615} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.086852] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1040.087818] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69fab483-cfeb-4922-b91f-330a8fee59bb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.093491] env[61987]: INFO nova.compute.manager [-] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Took 1.49 seconds to deallocate network for instance. [ 1040.101242] env[61987]: DEBUG nova.network.neutron [req-95731062-4ce5-41e5-a9d4-997767d44e1d req-6b300f60-adf1-44ee-bd96-efe597759e6e service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Updated VIF entry in instance network info cache for port d96dede6-c439-4a14-a4d1-62c44e03ea11. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1040.101655] env[61987]: DEBUG nova.network.neutron [req-95731062-4ce5-41e5-a9d4-997767d44e1d req-6b300f60-adf1-44ee-bd96-efe597759e6e service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Updating instance_info_cache with network_info: [{"id": "d96dede6-c439-4a14-a4d1-62c44e03ea11", "address": "fa:16:3e:d2:e0:b1", "network": {"id": "4ed4fbdc-05b0-44f8-a3d5-b7b288abd5e0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2067313775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a234a62da0e498fbe613fbcaaec3201", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd96dede6-c4", "ovs_interfaceid": "d96dede6-c439-4a14-a4d1-62c44e03ea11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.123307] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] b172195a-79d2-4f24-b9df-320a07de965e/b172195a-79d2-4f24-b9df-320a07de965e.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1040.124641] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2b2bfd7-85b2-4eee-afbb-55f0882ddab7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.144543] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1040.144543] env[61987]: value = "task-1062271" [ 1040.144543] env[61987]: _type = "Task" [ 1040.144543] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.152451] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062271, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.303130] env[61987]: DEBUG nova.network.neutron [-] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.360389] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-32257785-61d6-40d3-af95-0ceda514053d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.369503] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d6a5f1-3562-439b-9f86-414789e50dfd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.400379] env[61987]: DEBUG nova.compute.manager [req-5bb2d1d0-cfcf-43b3-80c6-14207b954ca6 req-4f06c3dc-40ae-48f3-94f1-2ee931f8e005 service nova] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Detach interface failed, port_id=b5999bb0-480a-45bd-b07c-fbb91985421e, reason: Instance f3773cda-41e8-4804-9319-c7e458112ddf could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 1040.605133] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.605618] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.605848] env[61987]: DEBUG nova.objects.instance [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lazy-loading 'resources' on Instance uuid dab6d2c5-8c04-4977-a409-b384037eabce {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.626033] env[61987]: DEBUG oslo_concurrency.lockutils [req-95731062-4ce5-41e5-a9d4-997767d44e1d req-6b300f60-adf1-44ee-bd96-efe597759e6e service nova] Releasing lock "refresh_cache-7002962c-76ed-4a8d-be2a-b2b118847fea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.656868] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062271, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.806153] env[61987]: INFO nova.compute.manager [-] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Took 1.65 seconds to deallocate network for instance. [ 1041.156837] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062271, 'name': ReconfigVM_Task, 'duration_secs': 0.971142} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.157464] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Reconfigured VM instance instance-0000006b to attach disk [datastore2] b172195a-79d2-4f24-b9df-320a07de965e/b172195a-79d2-4f24-b9df-320a07de965e.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1041.160723] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-58b868eb-5847-435f-a82c-80d8dc827bf4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.168996] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1041.168996] env[61987]: value = "task-1062272" [ 1041.168996] env[61987]: _type = "Task" [ 1041.168996] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.180379] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062272, 'name': Rename_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.294748] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e84385-5efa-4025-b76d-6ae90875a876 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.302687] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9a57449-be35-49c4-8d4d-9c341db5143c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.333553] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.334554] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19d7a414-9389-4eff-be66-840f85ab1c3e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.341780] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dedd1e14-2ddb-479a-ac9a-072e6f8bd367 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.356431] env[61987]: DEBUG nova.compute.provider_tree [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.679030] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062272, 'name': Rename_Task, 'duration_secs': 0.134393} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.679030] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1041.679196] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6cf9ea33-fee6-4ba4-af9f-4d3fff15bae7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.685450] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1041.685450] env[61987]: value = "task-1062273" [ 1041.685450] env[61987]: _type = "Task" [ 1041.685450] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.693028] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062273, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.859701] env[61987]: DEBUG nova.scheduler.client.report [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1041.945836] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Acquiring lock "24f8d398-beb0-4cc5-ada5-23f701adbe04" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.946095] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Lock "24f8d398-beb0-4cc5-ada5-23f701adbe04" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.195440] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062273, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.365764] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.760s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.368341] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.035s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.368621] env[61987]: DEBUG nova.objects.instance [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lazy-loading 'resources' on Instance uuid f3773cda-41e8-4804-9319-c7e458112ddf {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.392944] env[61987]: INFO nova.scheduler.client.report [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Deleted allocations for instance dab6d2c5-8c04-4977-a409-b384037eabce [ 1042.448927] env[61987]: DEBUG nova.compute.manager [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1042.697856] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062273, 'name': PowerOnVM_Task, 'duration_secs': 0.796374} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.698144] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1042.698400] env[61987]: DEBUG nova.compute.manager [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1042.699124] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e7401d-5e2b-4f44-ae18-e01ae7068066 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.900952] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5673dbc4-546d-49d3-b736-5035bb641a4f tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lock "dab6d2c5-8c04-4977-a409-b384037eabce" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.968s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.966554] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.050652] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e6a0855-f7a8-4305-9db9-87b9bb1f0c42 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.060293] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea36ae3e-b7d4-43b8-a6da-d4db09a949b0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.091485] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ea43a9e-341e-4c6c-b849-31f7b08949a9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.099103] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc53fae6-388a-41e0-a28b-c344a69ee63a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.112579] env[61987]: DEBUG nova.compute.provider_tree [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1043.212685] env[61987]: INFO nova.compute.manager [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] bringing vm to original state: 'stopped' [ 1043.554727] env[61987]: DEBUG oslo_concurrency.lockutils [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "interface-7be2179d-68f7-4103-a8c1-b6e5b88b8706-b67c9377-1e7d-4264-ae4c-dee82f96be5b" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.555131] env[61987]: DEBUG oslo_concurrency.lockutils [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "interface-7be2179d-68f7-4103-a8c1-b6e5b88b8706-b67c9377-1e7d-4264-ae4c-dee82f96be5b" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.555589] env[61987]: DEBUG nova.objects.instance [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lazy-loading 'flavor' on Instance uuid 7be2179d-68f7-4103-a8c1-b6e5b88b8706 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.615232] env[61987]: DEBUG nova.scheduler.client.report [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1043.894493] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquiring lock "e6236bf1-4068-4a6e-9787-cd812381e122" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.894846] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lock "e6236bf1-4068-4a6e-9787-cd812381e122" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.895088] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquiring lock "e6236bf1-4068-4a6e-9787-cd812381e122-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.895291] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lock "e6236bf1-4068-4a6e-9787-cd812381e122-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.895522] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lock "e6236bf1-4068-4a6e-9787-cd812381e122-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.897766] env[61987]: INFO nova.compute.manager [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Terminating instance [ 1044.121947] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.753s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.124119] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.158s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.126510] env[61987]: INFO nova.compute.claims [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1044.146149] env[61987]: INFO nova.scheduler.client.report [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Deleted allocations for instance f3773cda-41e8-4804-9319-c7e458112ddf [ 1044.168281] env[61987]: DEBUG nova.objects.instance [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lazy-loading 'pci_requests' on Instance uuid 7be2179d-68f7-4103-a8c1-b6e5b88b8706 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.219656] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "b172195a-79d2-4f24-b9df-320a07de965e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.220199] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "b172195a-79d2-4f24-b9df-320a07de965e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.220596] env[61987]: DEBUG nova.compute.manager [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1044.221577] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c50847dd-e6f4-4f8f-b19b-e74082c969d5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.228746] env[61987]: DEBUG nova.compute.manager [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61987) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1044.401433] env[61987]: DEBUG nova.compute.manager [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1044.401686] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1044.402658] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c9757d-f862-44f4-b1b1-79f55e198d72 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.410758] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1044.411016] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ce314d84-5585-4809-875f-a66cdc165afe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.417592] env[61987]: DEBUG oslo_vmware.api [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1044.417592] env[61987]: value = "task-1062274" [ 1044.417592] env[61987]: _type = "Task" [ 1044.417592] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.425819] env[61987]: DEBUG oslo_vmware.api [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062274, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.654364] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f044bd30-a634-4b6b-bbe9-32fe394c6754 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "f3773cda-41e8-4804-9319-c7e458112ddf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.127s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.670365] env[61987]: DEBUG nova.objects.base [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Object Instance<7be2179d-68f7-4103-a8c1-b6e5b88b8706> lazy-loaded attributes: flavor,pci_requests {{(pid=61987) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1044.670617] env[61987]: DEBUG nova.network.neutron [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1044.734315] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1044.734766] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-057435b1-52b2-4dae-8303-af8008c5c556 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.744914] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1044.744914] env[61987]: value = "task-1062275" [ 1044.744914] env[61987]: _type = "Task" [ 1044.744914] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.753510] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062275, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.755037] env[61987]: DEBUG nova.policy [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '697f9b1e3b3840369549a4f1ae3173ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd2afdc9bddba4ccab2b8fdb1f31da2b6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 1044.931396] env[61987]: DEBUG oslo_vmware.api [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062274, 'name': PowerOffVM_Task, 'duration_secs': 0.230019} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.931396] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1044.931396] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1044.931396] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fd8489f1-a7ff-42f0-bc35-0fc4f5d54702 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.990457] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1044.991486] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1044.991486] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Deleting the datastore file [datastore2] e6236bf1-4068-4a6e-9787-cd812381e122 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1044.992811] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9252ef26-1213-4269-8a81-c14d986fe198 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.001019] env[61987]: DEBUG oslo_vmware.api [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for the task: (returnval){ [ 1045.001019] env[61987]: value = "task-1062277" [ 1045.001019] env[61987]: _type = "Task" [ 1045.001019] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.006893] env[61987]: DEBUG oslo_vmware.api [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062277, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.257327] env[61987]: DEBUG oslo_vmware.api [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062275, 'name': PowerOffVM_Task, 'duration_secs': 0.34772} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.257649] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1045.258178] env[61987]: DEBUG nova.compute.manager [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1045.258787] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c79fbf0-ee54-4d1d-a922-537fe6054218 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.300952] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3601dd25-ceae-44ea-9c6b-ba38d00a7ea3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.308461] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a49bcee7-ed70-489e-9d4b-8e6a399b91b3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.338846] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a399ebce-d2a8-4d15-98f9-368f8805dab4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.345666] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd5eef07-7fb8-4eeb-b7ff-b7430cb80de5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.358608] env[61987]: DEBUG nova.compute.provider_tree [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.508287] env[61987]: DEBUG oslo_vmware.api [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Task: {'id': task-1062277, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.247906} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.508525] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1045.508698] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1045.508958] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1045.509135] env[61987]: INFO nova.compute.manager [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1045.509319] env[61987]: DEBUG oslo.service.loopingcall [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1045.509521] env[61987]: DEBUG nova.compute.manager [-] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1045.509611] env[61987]: DEBUG nova.network.neutron [-] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1045.778067] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "b172195a-79d2-4f24-b9df-320a07de965e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.558s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.861947] env[61987]: DEBUG nova.scheduler.client.report [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1045.884864] env[61987]: DEBUG nova.compute.manager [req-e7035f07-88ec-45de-bb63-ee129019d817 req-1f9ad56f-4f32-49e9-89cc-7fd7e3421804 service nova] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Received event network-vif-deleted-e9aa9f0a-f5e5-4d71-bddd-c817a7b4baf1 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1045.885142] env[61987]: INFO nova.compute.manager [req-e7035f07-88ec-45de-bb63-ee129019d817 req-1f9ad56f-4f32-49e9-89cc-7fd7e3421804 service nova] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Neutron deleted interface e9aa9f0a-f5e5-4d71-bddd-c817a7b4baf1; detaching it from the instance and deleting it from the info cache [ 1045.885336] env[61987]: DEBUG nova.network.neutron [req-e7035f07-88ec-45de-bb63-ee129019d817 req-1f9ad56f-4f32-49e9-89cc-7fd7e3421804 service nova] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.287124] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.338606] env[61987]: DEBUG nova.network.neutron [-] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.367248] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.243s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.367764] env[61987]: DEBUG nova.compute.manager [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1046.374017] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.084s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.374017] env[61987]: DEBUG nova.objects.instance [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61987) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1046.387959] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bb0d9630-d360-4fcc-9ff3-f6d6af3f6ffd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.399221] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fea3234-7439-4f4f-ae5e-2eb18bccb2fd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.427932] env[61987]: DEBUG nova.compute.manager [req-e7035f07-88ec-45de-bb63-ee129019d817 req-1f9ad56f-4f32-49e9-89cc-7fd7e3421804 service nova] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Detach interface failed, port_id=e9aa9f0a-f5e5-4d71-bddd-c817a7b4baf1, reason: Instance e6236bf1-4068-4a6e-9787-cd812381e122 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 1046.498114] env[61987]: DEBUG nova.network.neutron [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Successfully updated port: b67c9377-1e7d-4264-ae4c-dee82f96be5b {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1046.758995] env[61987]: DEBUG nova.compute.manager [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Stashing vm_state: active {{(pid=61987) _prep_resize /opt/stack/nova/nova/compute/manager.py:6090}} [ 1046.842073] env[61987]: INFO nova.compute.manager [-] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Took 1.33 seconds to deallocate network for instance. [ 1046.874256] env[61987]: DEBUG nova.compute.utils [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1046.878531] env[61987]: DEBUG nova.compute.manager [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1046.878618] env[61987]: DEBUG nova.network.neutron [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1046.923879] env[61987]: DEBUG nova.policy [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6de00ca7185f42929301ef66b1fb5573', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '24a539800b37438ca844c667731b5834', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 1047.005432] env[61987]: DEBUG oslo_concurrency.lockutils [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.005432] env[61987]: DEBUG oslo_concurrency.lockutils [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.005432] env[61987]: DEBUG nova.network.neutron [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1047.114509] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "b172195a-79d2-4f24-b9df-320a07de965e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.114801] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "b172195a-79d2-4f24-b9df-320a07de965e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.115032] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "b172195a-79d2-4f24-b9df-320a07de965e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.115233] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "b172195a-79d2-4f24-b9df-320a07de965e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.115412] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "b172195a-79d2-4f24-b9df-320a07de965e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.118492] env[61987]: INFO nova.compute.manager [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Terminating instance [ 1047.169780] env[61987]: DEBUG nova.network.neutron [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Successfully created port: ca34ba16-38ec-417f-b95f-ec6e425e5f61 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1047.283396] env[61987]: DEBUG oslo_concurrency.lockutils [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.349722] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.382652] env[61987]: DEBUG nova.compute.manager [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1047.386283] env[61987]: DEBUG oslo_concurrency.lockutils [None req-8e712520-7baf-4b90-a7df-ef9d973c4a01 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.387273] env[61987]: DEBUG oslo_concurrency.lockutils [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.104s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.536939] env[61987]: WARNING nova.network.neutron [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] 257456d2-c9db-4303-bdfd-035bbb123901 already exists in list: networks containing: ['257456d2-c9db-4303-bdfd-035bbb123901']. ignoring it [ 1047.622128] env[61987]: DEBUG nova.compute.manager [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1047.622372] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1047.623276] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bf2b627-7f0d-4dbe-abc4-cfe0fadaba30 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.630816] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1047.631088] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-81463f94-3c0d-42f2-b046-aa908e1ff681 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.696797] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1047.697063] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1047.697257] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Deleting the datastore file [datastore2] b172195a-79d2-4f24-b9df-320a07de965e {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1047.697515] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d40048e4-0085-4ffe-9e37-0d3fd72d27db {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.703402] env[61987]: DEBUG oslo_vmware.api [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1047.703402] env[61987]: value = "task-1062279" [ 1047.703402] env[61987]: _type = "Task" [ 1047.703402] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.710812] env[61987]: DEBUG oslo_vmware.api [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062279, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.849417] env[61987]: DEBUG nova.network.neutron [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Updating instance_info_cache with network_info: [{"id": "fe549122-49ad-4628-b106-216d787be2d4", "address": "fa:16:3e:76:0c:4d", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe549122-49", "ovs_interfaceid": "fe549122-49ad-4628-b106-216d787be2d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b67c9377-1e7d-4264-ae4c-dee82f96be5b", "address": "fa:16:3e:9e:56:e8", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb67c9377-1e", "ovs_interfaceid": "b67c9377-1e7d-4264-ae4c-dee82f96be5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.895511] env[61987]: INFO nova.compute.claims [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1047.995745] env[61987]: DEBUG nova.compute.manager [req-5b0a4ed2-c1ca-482d-b193-21065fa71740 req-a68760ef-68e6-4fcb-bd1f-cf903972c1df service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Received event network-vif-plugged-b67c9377-1e7d-4264-ae4c-dee82f96be5b {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1047.996020] env[61987]: DEBUG oslo_concurrency.lockutils [req-5b0a4ed2-c1ca-482d-b193-21065fa71740 req-a68760ef-68e6-4fcb-bd1f-cf903972c1df service nova] Acquiring lock "7be2179d-68f7-4103-a8c1-b6e5b88b8706-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.996257] env[61987]: DEBUG oslo_concurrency.lockutils [req-5b0a4ed2-c1ca-482d-b193-21065fa71740 req-a68760ef-68e6-4fcb-bd1f-cf903972c1df service nova] Lock "7be2179d-68f7-4103-a8c1-b6e5b88b8706-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.996438] env[61987]: DEBUG oslo_concurrency.lockutils [req-5b0a4ed2-c1ca-482d-b193-21065fa71740 req-a68760ef-68e6-4fcb-bd1f-cf903972c1df service nova] Lock "7be2179d-68f7-4103-a8c1-b6e5b88b8706-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.996649] env[61987]: DEBUG nova.compute.manager [req-5b0a4ed2-c1ca-482d-b193-21065fa71740 req-a68760ef-68e6-4fcb-bd1f-cf903972c1df service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] No waiting events found dispatching network-vif-plugged-b67c9377-1e7d-4264-ae4c-dee82f96be5b {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1047.996841] env[61987]: WARNING nova.compute.manager [req-5b0a4ed2-c1ca-482d-b193-21065fa71740 req-a68760ef-68e6-4fcb-bd1f-cf903972c1df service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Received unexpected event network-vif-plugged-b67c9377-1e7d-4264-ae4c-dee82f96be5b for instance with vm_state active and task_state None. [ 1047.997024] env[61987]: DEBUG nova.compute.manager [req-5b0a4ed2-c1ca-482d-b193-21065fa71740 req-a68760ef-68e6-4fcb-bd1f-cf903972c1df service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Received event network-changed-b67c9377-1e7d-4264-ae4c-dee82f96be5b {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1047.997187] env[61987]: DEBUG nova.compute.manager [req-5b0a4ed2-c1ca-482d-b193-21065fa71740 req-a68760ef-68e6-4fcb-bd1f-cf903972c1df service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Refreshing instance network info cache due to event network-changed-b67c9377-1e7d-4264-ae4c-dee82f96be5b. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1047.997364] env[61987]: DEBUG oslo_concurrency.lockutils [req-5b0a4ed2-c1ca-482d-b193-21065fa71740 req-a68760ef-68e6-4fcb-bd1f-cf903972c1df service nova] Acquiring lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.212930] env[61987]: DEBUG oslo_vmware.api [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062279, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149214} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.213234] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1048.213451] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1048.213693] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1048.213886] env[61987]: INFO nova.compute.manager [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1048.214143] env[61987]: DEBUG oslo.service.loopingcall [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1048.214344] env[61987]: DEBUG nova.compute.manager [-] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1048.214436] env[61987]: DEBUG nova.network.neutron [-] [instance: b172195a-79d2-4f24-b9df-320a07de965e] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1048.352177] env[61987]: DEBUG oslo_concurrency.lockutils [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.352983] env[61987]: DEBUG oslo_concurrency.lockutils [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.353341] env[61987]: DEBUG oslo_concurrency.lockutils [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.353680] env[61987]: DEBUG oslo_concurrency.lockutils [req-5b0a4ed2-c1ca-482d-b193-21065fa71740 req-a68760ef-68e6-4fcb-bd1f-cf903972c1df service nova] Acquired lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.353897] env[61987]: DEBUG nova.network.neutron [req-5b0a4ed2-c1ca-482d-b193-21065fa71740 req-a68760ef-68e6-4fcb-bd1f-cf903972c1df service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Refreshing network info cache for port b67c9377-1e7d-4264-ae4c-dee82f96be5b {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1048.355900] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f427d05-38a9-4611-aa2c-38dedab6636b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.373212] env[61987]: DEBUG nova.virt.hardware [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1048.373463] env[61987]: DEBUG nova.virt.hardware [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1048.373630] env[61987]: DEBUG nova.virt.hardware [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1048.373841] env[61987]: DEBUG nova.virt.hardware [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1048.374012] env[61987]: DEBUG nova.virt.hardware [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1048.374185] env[61987]: DEBUG nova.virt.hardware [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1048.374397] env[61987]: DEBUG nova.virt.hardware [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1048.374579] env[61987]: DEBUG nova.virt.hardware [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1048.374762] env[61987]: DEBUG nova.virt.hardware [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1048.374995] env[61987]: DEBUG nova.virt.hardware [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1048.375122] env[61987]: DEBUG nova.virt.hardware [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1048.381444] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Reconfiguring VM to attach interface {{(pid=61987) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1048.382333] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9343f49-a622-4c8d-a729-45d8a402cad4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.400487] env[61987]: DEBUG oslo_vmware.api [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1048.400487] env[61987]: value = "task-1062280" [ 1048.400487] env[61987]: _type = "Task" [ 1048.400487] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.406009] env[61987]: INFO nova.compute.resource_tracker [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updating resource usage from migration c7b03f3d-2ed7-4f0f-89bf-a42b9a6f5af6 [ 1048.409268] env[61987]: DEBUG nova.compute.manager [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1048.419965] env[61987]: DEBUG oslo_vmware.api [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062280, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.455102] env[61987]: DEBUG nova.virt.hardware [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1048.455227] env[61987]: DEBUG nova.virt.hardware [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1048.455347] env[61987]: DEBUG nova.virt.hardware [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1048.455555] env[61987]: DEBUG nova.virt.hardware [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1048.455702] env[61987]: DEBUG nova.virt.hardware [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1048.455896] env[61987]: DEBUG nova.virt.hardware [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1048.456142] env[61987]: DEBUG nova.virt.hardware [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1048.456314] env[61987]: DEBUG nova.virt.hardware [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1048.456518] env[61987]: DEBUG nova.virt.hardware [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1048.456662] env[61987]: DEBUG nova.virt.hardware [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1048.456840] env[61987]: DEBUG nova.virt.hardware [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1048.457748] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eee1ce6-e622-4011-9019-47761248a770 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.468189] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-740a9a6f-4457-4486-b13d-92c9526d30f0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.580078] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f62d0af1-91e6-44ac-bc1f-eab476856e18 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.586836] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7416f0f-ea46-4cfe-9cac-a87ac7794eb8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.616804] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58585e8e-8e3d-46bc-a23e-136ad8d4acc1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.624395] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c67f1a-c8f5-454b-8afe-c63593db396f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.638237] env[61987]: DEBUG nova.compute.provider_tree [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1048.653074] env[61987]: DEBUG nova.network.neutron [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Successfully updated port: ca34ba16-38ec-417f-b95f-ec6e425e5f61 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1048.910564] env[61987]: DEBUG oslo_vmware.api [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062280, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.914099] env[61987]: DEBUG nova.network.neutron [-] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.141133] env[61987]: DEBUG nova.scheduler.client.report [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1049.145054] env[61987]: DEBUG nova.network.neutron [req-5b0a4ed2-c1ca-482d-b193-21065fa71740 req-a68760ef-68e6-4fcb-bd1f-cf903972c1df service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Updated VIF entry in instance network info cache for port b67c9377-1e7d-4264-ae4c-dee82f96be5b. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1049.145471] env[61987]: DEBUG nova.network.neutron [req-5b0a4ed2-c1ca-482d-b193-21065fa71740 req-a68760ef-68e6-4fcb-bd1f-cf903972c1df service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Updating instance_info_cache with network_info: [{"id": "fe549122-49ad-4628-b106-216d787be2d4", "address": "fa:16:3e:76:0c:4d", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe549122-49", "ovs_interfaceid": "fe549122-49ad-4628-b106-216d787be2d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b67c9377-1e7d-4264-ae4c-dee82f96be5b", "address": "fa:16:3e:9e:56:e8", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb67c9377-1e", "ovs_interfaceid": "b67c9377-1e7d-4264-ae4c-dee82f96be5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.155078] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Acquiring lock "refresh_cache-24f8d398-beb0-4cc5-ada5-23f701adbe04" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.155213] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Acquired lock "refresh_cache-24f8d398-beb0-4cc5-ada5-23f701adbe04" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.155352] env[61987]: DEBUG nova.network.neutron [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1049.409708] env[61987]: DEBUG oslo_vmware.api [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062280, 'name': ReconfigVM_Task, 'duration_secs': 0.603115} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.410268] env[61987]: DEBUG oslo_concurrency.lockutils [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.410497] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Reconfigured VM to attach interface {{(pid=61987) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1049.418648] env[61987]: INFO nova.compute.manager [-] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Took 1.20 seconds to deallocate network for instance. [ 1049.648021] env[61987]: DEBUG oslo_concurrency.lockutils [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.261s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.648257] env[61987]: INFO nova.compute.manager [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Migrating [ 1049.654684] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.305s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.654968] env[61987]: DEBUG nova.objects.instance [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lazy-loading 'resources' on Instance uuid e6236bf1-4068-4a6e-9787-cd812381e122 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.656331] env[61987]: DEBUG oslo_concurrency.lockutils [req-5b0a4ed2-c1ca-482d-b193-21065fa71740 req-a68760ef-68e6-4fcb-bd1f-cf903972c1df service nova] Releasing lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.665476] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1049.666334] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1049.666483] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Starting heal instance info cache {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10401}} [ 1049.691783] env[61987]: DEBUG nova.network.neutron [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1049.811611] env[61987]: DEBUG nova.network.neutron [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Updating instance_info_cache with network_info: [{"id": "ca34ba16-38ec-417f-b95f-ec6e425e5f61", "address": "fa:16:3e:4c:c1:f3", "network": {"id": "632be0c3-c9b7-4cf8-b92f-9a58c0da2be6", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1857904253-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24a539800b37438ca844c667731b5834", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca34ba16-38", "ovs_interfaceid": "ca34ba16-38ec-417f-b95f-ec6e425e5f61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.915264] env[61987]: DEBUG oslo_concurrency.lockutils [None req-07f4844f-ee36-4f13-9cfd-ca1d1ab0d8ac tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "interface-7be2179d-68f7-4103-a8c1-b6e5b88b8706-b67c9377-1e7d-4264-ae4c-dee82f96be5b" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.360s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.924403] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.025136] env[61987]: DEBUG nova.compute.manager [req-20552b0f-f341-401d-a24b-0de8cecdc56f req-c8737ffd-696c-4230-a51a-34f418aebed9 service nova] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Received event network-vif-deleted-2a9d7219-a935-4e81-9278-7abce87993fe {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1050.025374] env[61987]: DEBUG nova.compute.manager [req-20552b0f-f341-401d-a24b-0de8cecdc56f req-c8737ffd-696c-4230-a51a-34f418aebed9 service nova] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Received event network-vif-plugged-ca34ba16-38ec-417f-b95f-ec6e425e5f61 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1050.025597] env[61987]: DEBUG oslo_concurrency.lockutils [req-20552b0f-f341-401d-a24b-0de8cecdc56f req-c8737ffd-696c-4230-a51a-34f418aebed9 service nova] Acquiring lock "24f8d398-beb0-4cc5-ada5-23f701adbe04-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.025823] env[61987]: DEBUG oslo_concurrency.lockutils [req-20552b0f-f341-401d-a24b-0de8cecdc56f req-c8737ffd-696c-4230-a51a-34f418aebed9 service nova] Lock "24f8d398-beb0-4cc5-ada5-23f701adbe04-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.026008] env[61987]: DEBUG oslo_concurrency.lockutils [req-20552b0f-f341-401d-a24b-0de8cecdc56f req-c8737ffd-696c-4230-a51a-34f418aebed9 service nova] Lock "24f8d398-beb0-4cc5-ada5-23f701adbe04-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.026191] env[61987]: DEBUG nova.compute.manager [req-20552b0f-f341-401d-a24b-0de8cecdc56f req-c8737ffd-696c-4230-a51a-34f418aebed9 service nova] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] No waiting events found dispatching network-vif-plugged-ca34ba16-38ec-417f-b95f-ec6e425e5f61 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1050.026365] env[61987]: WARNING nova.compute.manager [req-20552b0f-f341-401d-a24b-0de8cecdc56f req-c8737ffd-696c-4230-a51a-34f418aebed9 service nova] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Received unexpected event network-vif-plugged-ca34ba16-38ec-417f-b95f-ec6e425e5f61 for instance with vm_state building and task_state spawning. [ 1050.026548] env[61987]: DEBUG nova.compute.manager [req-20552b0f-f341-401d-a24b-0de8cecdc56f req-c8737ffd-696c-4230-a51a-34f418aebed9 service nova] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Received event network-changed-ca34ba16-38ec-417f-b95f-ec6e425e5f61 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1050.026701] env[61987]: DEBUG nova.compute.manager [req-20552b0f-f341-401d-a24b-0de8cecdc56f req-c8737ffd-696c-4230-a51a-34f418aebed9 service nova] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Refreshing instance network info cache due to event network-changed-ca34ba16-38ec-417f-b95f-ec6e425e5f61. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1050.026870] env[61987]: DEBUG oslo_concurrency.lockutils [req-20552b0f-f341-401d-a24b-0de8cecdc56f req-c8737ffd-696c-4230-a51a-34f418aebed9 service nova] Acquiring lock "refresh_cache-24f8d398-beb0-4cc5-ada5-23f701adbe04" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.166851] env[61987]: DEBUG oslo_concurrency.lockutils [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.167100] env[61987]: DEBUG oslo_concurrency.lockutils [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.167677] env[61987]: DEBUG nova.network.neutron [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1050.170483] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.310959] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7771107-febc-459f-9bc2-f01134037602 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.313796] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Releasing lock "refresh_cache-24f8d398-beb0-4cc5-ada5-23f701adbe04" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.314103] env[61987]: DEBUG nova.compute.manager [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Instance network_info: |[{"id": "ca34ba16-38ec-417f-b95f-ec6e425e5f61", "address": "fa:16:3e:4c:c1:f3", "network": {"id": "632be0c3-c9b7-4cf8-b92f-9a58c0da2be6", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1857904253-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24a539800b37438ca844c667731b5834", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca34ba16-38", "ovs_interfaceid": "ca34ba16-38ec-417f-b95f-ec6e425e5f61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1050.314397] env[61987]: DEBUG oslo_concurrency.lockutils [req-20552b0f-f341-401d-a24b-0de8cecdc56f req-c8737ffd-696c-4230-a51a-34f418aebed9 service nova] Acquired lock "refresh_cache-24f8d398-beb0-4cc5-ada5-23f701adbe04" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.314609] env[61987]: DEBUG nova.network.neutron [req-20552b0f-f341-401d-a24b-0de8cecdc56f req-c8737ffd-696c-4230-a51a-34f418aebed9 service nova] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Refreshing network info cache for port ca34ba16-38ec-417f-b95f-ec6e425e5f61 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1050.315960] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4c:c1:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd733acc2-07d0-479e-918c-ec8a21925389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ca34ba16-38ec-417f-b95f-ec6e425e5f61', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1050.323194] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Creating folder: Project (24a539800b37438ca844c667731b5834). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1050.324583] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1c626ab6-7536-4c6c-ac2a-722b52079222 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.330543] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a43504-0962-4031-a3d6-3736e161f7cd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.362200] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c1734e-ddb1-451c-a80a-f056e006ad74 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.364873] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Created folder: Project (24a539800b37438ca844c667731b5834) in parent group-v234219. [ 1050.365075] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Creating folder: Instances. Parent ref: group-v234427. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1050.365303] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9134175c-cf2a-4315-b941-2ddbcf811f7c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.371924] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a29fc183-5a9c-44f1-9926-08b41823f2ad {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.376558] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Created folder: Instances in parent group-v234427. [ 1050.376806] env[61987]: DEBUG oslo.service.loopingcall [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1050.377326] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1050.377559] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-47524a4c-972e-4e66-93e9-60b8d040b3dc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.399228] env[61987]: DEBUG nova.compute.provider_tree [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1050.404760] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1050.404760] env[61987]: value = "task-1062283" [ 1050.404760] env[61987]: _type = "Task" [ 1050.404760] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.412459] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062283, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.901639] env[61987]: DEBUG nova.scheduler.client.report [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1050.922588] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062283, 'name': CreateVM_Task, 'duration_secs': 0.30423} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.922847] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1050.924067] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.924261] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.926412] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1050.927446] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-896a62df-2e83-4f11-b9e7-ceb577f3bd60 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.934455] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Waiting for the task: (returnval){ [ 1050.934455] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52a97d45-a056-c036-4ba6-5acf29070e5f" [ 1050.934455] env[61987]: _type = "Task" [ 1050.934455] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.942146] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52a97d45-a056-c036-4ba6-5acf29070e5f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.020189] env[61987]: DEBUG nova.network.neutron [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updating instance_info_cache with network_info: [{"id": "785edefc-5e84-4a34-97e1-b7c1a08f218e", "address": "fa:16:3e:bf:8b:d7", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap785edefc-5e", "ovs_interfaceid": "785edefc-5e84-4a34-97e1-b7c1a08f218e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.100324] env[61987]: DEBUG nova.network.neutron [req-20552b0f-f341-401d-a24b-0de8cecdc56f req-c8737ffd-696c-4230-a51a-34f418aebed9 service nova] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Updated VIF entry in instance network info cache for port ca34ba16-38ec-417f-b95f-ec6e425e5f61. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1051.100679] env[61987]: DEBUG nova.network.neutron [req-20552b0f-f341-401d-a24b-0de8cecdc56f req-c8737ffd-696c-4230-a51a-34f418aebed9 service nova] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Updating instance_info_cache with network_info: [{"id": "ca34ba16-38ec-417f-b95f-ec6e425e5f61", "address": "fa:16:3e:4c:c1:f3", "network": {"id": "632be0c3-c9b7-4cf8-b92f-9a58c0da2be6", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1857904253-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "24a539800b37438ca844c667731b5834", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d733acc2-07d0-479e-918c-ec8a21925389", "external-id": "nsx-vlan-transportzone-459", "segmentation_id": 459, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca34ba16-38", "ovs_interfaceid": "ca34ba16-38ec-417f-b95f-ec6e425e5f61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.393196] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "interface-7be2179d-68f7-4103-a8c1-b6e5b88b8706-b67c9377-1e7d-4264-ae4c-dee82f96be5b" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.393493] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "interface-7be2179d-68f7-4103-a8c1-b6e5b88b8706-b67c9377-1e7d-4264-ae4c-dee82f96be5b" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.405635] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.751s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.407788] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.483s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.408620] env[61987]: DEBUG nova.objects.instance [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lazy-loading 'resources' on Instance uuid b172195a-79d2-4f24-b9df-320a07de965e {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1051.425928] env[61987]: INFO nova.scheduler.client.report [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Deleted allocations for instance e6236bf1-4068-4a6e-9787-cd812381e122 [ 1051.444915] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52a97d45-a056-c036-4ba6-5acf29070e5f, 'name': SearchDatastore_Task, 'duration_secs': 0.009319} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.445273] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.445519] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1051.445796] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.445960] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.446162] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1051.446433] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-38268890-d480-47ff-95f2-e302c8980758 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.455419] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1051.455419] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1051.455419] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c08bf622-e6b2-4f99-903b-a6fd4fc4db21 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.460586] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Waiting for the task: (returnval){ [ 1051.460586] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52ca3eeb-e3f6-f559-6508-46f1b0480161" [ 1051.460586] env[61987]: _type = "Task" [ 1051.460586] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.468521] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52ca3eeb-e3f6-f559-6508-46f1b0480161, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.522902] env[61987]: DEBUG oslo_concurrency.lockutils [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.524410] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquired lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.524593] env[61987]: DEBUG nova.network.neutron [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Forcefully refreshing network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1051.604120] env[61987]: DEBUG oslo_concurrency.lockutils [req-20552b0f-f341-401d-a24b-0de8cecdc56f req-c8737ffd-696c-4230-a51a-34f418aebed9 service nova] Releasing lock "refresh_cache-24f8d398-beb0-4cc5-ada5-23f701adbe04" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.896626] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.896884] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.898218] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2658e317-fad9-4aae-aa6c-ce48771af1cc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.918559] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-087fc7bb-0443-4f63-aeba-dd9bf09b0183 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.946461] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Reconfiguring VM to detach interface {{(pid=61987) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1051.949389] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a53e741b-612f-476c-b75e-92757de91220 tempest-ServerRescueTestJSON-953534140 tempest-ServerRescueTestJSON-953534140-project-member] Lock "e6236bf1-4068-4a6e-9787-cd812381e122" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.055s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.950574] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ef01904-9b18-432e-935f-25f82988374f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.975060] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52ca3eeb-e3f6-f559-6508-46f1b0480161, 'name': SearchDatastore_Task, 'duration_secs': 0.010364} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.979522] env[61987]: DEBUG oslo_vmware.api [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1051.979522] env[61987]: value = "task-1062284" [ 1051.979522] env[61987]: _type = "Task" [ 1051.979522] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.979940] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-022aef72-f2d3-461b-84b5-dc2165392b03 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.988200] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Waiting for the task: (returnval){ [ 1051.988200] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]526a57a8-1d67-f9eb-849e-30d0a744e3bc" [ 1051.988200] env[61987]: _type = "Task" [ 1051.988200] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.994307] env[61987]: DEBUG oslo_vmware.api [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062284, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.003551] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]526a57a8-1d67-f9eb-849e-30d0a744e3bc, 'name': SearchDatastore_Task, 'duration_secs': 0.009784} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.006351] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1052.006683] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 24f8d398-beb0-4cc5-ada5-23f701adbe04/24f8d398-beb0-4cc5-ada5-23f701adbe04.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1052.007452] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a5e154d7-3f12-4332-b9d1-c64753b7725e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.013353] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Waiting for the task: (returnval){ [ 1052.013353] env[61987]: value = "task-1062285" [ 1052.013353] env[61987]: _type = "Task" [ 1052.013353] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.023261] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Task: {'id': task-1062285, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.100428] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bca8954b-66e6-4e2b-8278-700be9ae0c35 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.108176] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b2d5fd0-09c1-439c-a2eb-5b9eaaf92bc2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.138258] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b686d5e9-a7f2-465a-9b67-5eec23a44746 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.145138] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9d2bd84-1466-482b-88ca-f9c88c3dfe6f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.161887] env[61987]: DEBUG nova.compute.provider_tree [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1052.492959] env[61987]: DEBUG oslo_vmware.api [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062284, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.524877] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Task: {'id': task-1062285, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461481} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.524877] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 24f8d398-beb0-4cc5-ada5-23f701adbe04/24f8d398-beb0-4cc5-ada5-23f701adbe04.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1052.524877] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1052.524877] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e2f01f1b-aebb-42c8-b8a6-6f1b05800012 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.532027] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Waiting for the task: (returnval){ [ 1052.532027] env[61987]: value = "task-1062286" [ 1052.532027] env[61987]: _type = "Task" [ 1052.532027] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.540785] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Task: {'id': task-1062286, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.665335] env[61987]: DEBUG nova.scheduler.client.report [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1052.756826] env[61987]: DEBUG nova.network.neutron [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updating instance_info_cache with network_info: [{"id": "785edefc-5e84-4a34-97e1-b7c1a08f218e", "address": "fa:16:3e:bf:8b:d7", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap785edefc-5e", "ovs_interfaceid": "785edefc-5e84-4a34-97e1-b7c1a08f218e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.993067] env[61987]: DEBUG oslo_vmware.api [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062284, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.040685] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Task: {'id': task-1062286, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.05778} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.040946] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1053.041724] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31d550ec-84d7-43b4-8785-0951401616b3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.045552] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba725160-2bb5-4802-9dd1-cdd7c4da3956 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.071571] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 24f8d398-beb0-4cc5-ada5-23f701adbe04/24f8d398-beb0-4cc5-ada5-23f701adbe04.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1053.084437] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2dca9c57-3d09-4200-a8e2-43076e20f19a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.098939] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updating instance '69a22e40-d469-4500-926e-0a12a233f252' progress to 0 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1053.107772] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Waiting for the task: (returnval){ [ 1053.107772] env[61987]: value = "task-1062287" [ 1053.107772] env[61987]: _type = "Task" [ 1053.107772] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.115648] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Task: {'id': task-1062287, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.175018] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.764s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.204310] env[61987]: INFO nova.scheduler.client.report [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Deleted allocations for instance b172195a-79d2-4f24-b9df-320a07de965e [ 1053.259892] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Releasing lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.260167] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updated the network info_cache for instance {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10472}} [ 1053.260410] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1053.260589] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1053.260744] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1053.260898] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1053.261070] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1053.261231] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1053.261362] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61987) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11020}} [ 1053.261549] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1053.494527] env[61987]: DEBUG oslo_vmware.api [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062284, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.609093] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1053.609498] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-49ee2cfc-d4a6-46c9-b121-587a81d7c11f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.619949] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Task: {'id': task-1062287, 'name': ReconfigVM_Task, 'duration_secs': 0.2724} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.621194] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 24f8d398-beb0-4cc5-ada5-23f701adbe04/24f8d398-beb0-4cc5-ada5-23f701adbe04.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1053.623740] env[61987]: DEBUG oslo_vmware.api [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1053.623740] env[61987]: value = "task-1062288" [ 1053.623740] env[61987]: _type = "Task" [ 1053.623740] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.623740] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-54ab0de5-0117-4dbb-b480-7faf1b72603f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.631421] env[61987]: DEBUG oslo_vmware.api [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062288, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.632778] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Waiting for the task: (returnval){ [ 1053.632778] env[61987]: value = "task-1062289" [ 1053.632778] env[61987]: _type = "Task" [ 1053.632778] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.640833] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Task: {'id': task-1062289, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.712599] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b49cc5c3-7791-4d6d-a246-8b3b704f6637 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "b172195a-79d2-4f24-b9df-320a07de965e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.598s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.767022] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.767022] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.767022] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.767022] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61987) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1053.767022] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dfb9150-3a57-4f94-ab82-b6270e4de7ca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.778649] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34896bac-b02d-4287-ba60-11b81007018f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.793690] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-233b3d2b-2c1b-4eeb-a04f-ba6055bf2fb9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.802873] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c2c94a1-17f6-432c-9bf0-d218e5773fa5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.836217] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180224MB free_disk=179GB free_vcpus=48 pci_devices=None {{(pid=61987) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1053.836332] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.836528] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.994287] env[61987]: DEBUG oslo_vmware.api [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062284, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.132760] env[61987]: DEBUG oslo_vmware.api [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062288, 'name': PowerOffVM_Task, 'duration_secs': 0.205486} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.133149] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1054.133359] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updating instance '69a22e40-d469-4500-926e-0a12a233f252' progress to 17 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1054.144907] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Task: {'id': task-1062289, 'name': Rename_Task, 'duration_secs': 0.146606} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.145193] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1054.145436] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fbe4b998-3b02-4e41-a2fc-d91e40f6b0a0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.151383] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Waiting for the task: (returnval){ [ 1054.151383] env[61987]: value = "task-1062290" [ 1054.151383] env[61987]: _type = "Task" [ 1054.151383] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.158911] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Task: {'id': task-1062290, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.496041] env[61987]: DEBUG oslo_vmware.api [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062284, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.644444] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1054.645009] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1054.645247] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1054.645769] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1054.645998] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1054.646185] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1054.646409] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1054.646582] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1054.647130] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1054.647503] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1054.647829] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1054.654134] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d382339-9f16-48a0-8706-d38b6dd1ac84 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.677547] env[61987]: DEBUG oslo_vmware.api [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Task: {'id': task-1062290, 'name': PowerOnVM_Task, 'duration_secs': 0.450551} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.679066] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1054.679324] env[61987]: INFO nova.compute.manager [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Took 6.27 seconds to spawn the instance on the hypervisor. [ 1054.679389] env[61987]: DEBUG nova.compute.manager [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1054.679716] env[61987]: DEBUG oslo_vmware.api [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1054.679716] env[61987]: value = "task-1062291" [ 1054.679716] env[61987]: _type = "Task" [ 1054.679716] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.680424] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a67d042-e271-4944-9546-fd60c59be1d0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.695553] env[61987]: DEBUG oslo_vmware.api [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062291, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.846390] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Applying migration context for instance 69a22e40-d469-4500-926e-0a12a233f252 as it has an incoming, in-progress migration c7b03f3d-2ed7-4f0f-89bf-a42b9a6f5af6. Migration status is migrating {{(pid=61987) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1054.847753] env[61987]: INFO nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updating resource usage from migration c7b03f3d-2ed7-4f0f-89bf-a42b9a6f5af6 [ 1054.871290] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 4d15848d-5265-4897-bfa7-f3ad4c02ed83 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1054.871451] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1054.871581] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 473e0b73-6fcb-44c1-abf7-47d30c58af1b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1054.871703] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 7be2179d-68f7-4103-a8c1-b6e5b88b8706 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1054.871823] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 0cba5e51-7e0a-409b-a7b2-612ab8cf5689 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1054.872081] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance bf1123bf-80e6-4be3-804f-e13b906ee44d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1054.872081] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 7002962c-76ed-4a8d-be2a-b2b118847fea actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1054.872200] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 24f8d398-beb0-4cc5-ada5-23f701adbe04 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1054.872287] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Migration c7b03f3d-2ed7-4f0f-89bf-a42b9a6f5af6 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1054.872399] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 69a22e40-d469-4500-926e-0a12a233f252 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1054.872601] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1054.872741] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2496MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1054.995857] env[61987]: DEBUG oslo_vmware.api [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062284, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.010850] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-285a3261-b728-422f-bc35-b09007dee16b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.018581] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e23984a8-26e1-4bf0-8af3-80d7882a2c09 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.063408] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbdd5ed6-e454-4c7b-bfda-69cb4e0579b4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.073535] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ee253fb-54d2-48f6-b1d9-9b2817f9a611 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.087304] env[61987]: DEBUG nova.compute.provider_tree [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1055.192249] env[61987]: DEBUG oslo_vmware.api [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062291, 'name': ReconfigVM_Task, 'duration_secs': 0.302914} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.192579] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updating instance '69a22e40-d469-4500-926e-0a12a233f252' progress to 33 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1055.203440] env[61987]: INFO nova.compute.manager [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Took 12.25 seconds to build instance. [ 1055.496476] env[61987]: DEBUG oslo_vmware.api [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062284, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.590468] env[61987]: DEBUG nova.scheduler.client.report [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1055.625034] env[61987]: DEBUG oslo_concurrency.lockutils [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "76707afe-29ae-4257-a237-ba5a47b2511b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.625283] env[61987]: DEBUG oslo_concurrency.lockutils [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "76707afe-29ae-4257-a237-ba5a47b2511b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.699428] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1055.699761] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1055.699881] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1055.700070] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1055.700235] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1055.700394] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1055.700609] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1055.700780] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1055.700956] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1055.701141] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1055.701324] env[61987]: DEBUG nova.virt.hardware [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1055.709203] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Reconfiguring VM instance instance-0000004f to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1055.709942] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e07401a9-ce80-4108-923c-5f5840bf4f29 tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Lock "24f8d398-beb0-4cc5-ada5-23f701adbe04" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.764s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.710218] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36a387e3-8c48-49b4-867e-15a4b14a5bd2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.730121] env[61987]: DEBUG oslo_vmware.api [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1055.730121] env[61987]: value = "task-1062292" [ 1055.730121] env[61987]: _type = "Task" [ 1055.730121] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.738257] env[61987]: DEBUG oslo_vmware.api [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062292, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.911356] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Acquiring lock "24f8d398-beb0-4cc5-ada5-23f701adbe04" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.911645] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Lock "24f8d398-beb0-4cc5-ada5-23f701adbe04" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.911866] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Acquiring lock "24f8d398-beb0-4cc5-ada5-23f701adbe04-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.912074] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Lock "24f8d398-beb0-4cc5-ada5-23f701adbe04-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.912248] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Lock "24f8d398-beb0-4cc5-ada5-23f701adbe04-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.914409] env[61987]: INFO nova.compute.manager [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Terminating instance [ 1055.997061] env[61987]: DEBUG oslo_vmware.api [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062284, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.097134] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61987) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1056.097134] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.260s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.128062] env[61987]: DEBUG nova.compute.manager [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1056.240259] env[61987]: DEBUG oslo_vmware.api [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062292, 'name': ReconfigVM_Task, 'duration_secs': 0.159031} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.240531] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Reconfigured VM instance instance-0000004f to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1056.241307] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a6ad8c7-a3ae-4b7a-8430-ec17833571eb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.262879] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] 69a22e40-d469-4500-926e-0a12a233f252/69a22e40-d469-4500-926e-0a12a233f252.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1056.263124] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9493448-3d53-485f-a821-831c58aa5510 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.280220] env[61987]: DEBUG oslo_vmware.api [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1056.280220] env[61987]: value = "task-1062293" [ 1056.280220] env[61987]: _type = "Task" [ 1056.280220] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.287440] env[61987]: DEBUG oslo_vmware.api [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062293, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.417964] env[61987]: DEBUG nova.compute.manager [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1056.418302] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1056.419319] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f988895-d3b9-490a-9bab-8003e2b8b001 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.427601] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1056.427839] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-43fea9b7-8322-4f35-93f2-abbaa87b82be {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.433851] env[61987]: DEBUG oslo_vmware.api [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Waiting for the task: (returnval){ [ 1056.433851] env[61987]: value = "task-1062294" [ 1056.433851] env[61987]: _type = "Task" [ 1056.433851] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.441265] env[61987]: DEBUG oslo_vmware.api [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Task: {'id': task-1062294, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.499172] env[61987]: DEBUG oslo_vmware.api [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062284, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.649717] env[61987]: DEBUG oslo_concurrency.lockutils [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.649999] env[61987]: DEBUG oslo_concurrency.lockutils [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.651553] env[61987]: INFO nova.compute.claims [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1056.789734] env[61987]: DEBUG oslo_vmware.api [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062293, 'name': ReconfigVM_Task, 'duration_secs': 0.264702} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.789973] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Reconfigured VM instance instance-0000004f to attach disk [datastore2] 69a22e40-d469-4500-926e-0a12a233f252/69a22e40-d469-4500-926e-0a12a233f252.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1056.790278] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updating instance '69a22e40-d469-4500-926e-0a12a233f252' progress to 50 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1056.944934] env[61987]: DEBUG oslo_vmware.api [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Task: {'id': task-1062294, 'name': PowerOffVM_Task, 'duration_secs': 0.163632} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.945246] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1056.945422] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1056.945701] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af6fed04-fab6-4a85-a624-889d5ac90635 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.999415] env[61987]: DEBUG oslo_vmware.api [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062284, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.008174] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1057.008406] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1057.008603] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Deleting the datastore file [datastore1] 24f8d398-beb0-4cc5-ada5-23f701adbe04 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1057.008865] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2a2bbed9-64bc-4934-88c8-7b8bbc481221 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.014652] env[61987]: DEBUG oslo_vmware.api [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Waiting for the task: (returnval){ [ 1057.014652] env[61987]: value = "task-1062296" [ 1057.014652] env[61987]: _type = "Task" [ 1057.014652] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.022390] env[61987]: DEBUG oslo_vmware.api [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Task: {'id': task-1062296, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.298512] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2d6b7cb-0f36-4d3d-af03-49de445aae1c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.317239] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f10254e-6ae8-4554-bcc5-cda0b51a4ae8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.334053] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updating instance '69a22e40-d469-4500-926e-0a12a233f252' progress to 67 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1057.500292] env[61987]: DEBUG oslo_vmware.api [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062284, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.523819] env[61987]: DEBUG oslo_vmware.api [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Task: {'id': task-1062296, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13429} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.524066] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1057.524269] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1057.524454] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1057.524634] env[61987]: INFO nova.compute.manager [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1057.524894] env[61987]: DEBUG oslo.service.loopingcall [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1057.525106] env[61987]: DEBUG nova.compute.manager [-] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1057.525208] env[61987]: DEBUG nova.network.neutron [-] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1057.763170] env[61987]: DEBUG nova.compute.manager [req-49a2167e-d446-46f3-97bf-fc19a8cbf453 req-28798237-4984-4048-b6cd-b56d6d0904c9 service nova] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Received event network-vif-deleted-ca34ba16-38ec-417f-b95f-ec6e425e5f61 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1057.763398] env[61987]: INFO nova.compute.manager [req-49a2167e-d446-46f3-97bf-fc19a8cbf453 req-28798237-4984-4048-b6cd-b56d6d0904c9 service nova] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Neutron deleted interface ca34ba16-38ec-417f-b95f-ec6e425e5f61; detaching it from the instance and deleting it from the info cache [ 1057.763578] env[61987]: DEBUG nova.network.neutron [req-49a2167e-d446-46f3-97bf-fc19a8cbf453 req-28798237-4984-4048-b6cd-b56d6d0904c9 service nova] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.808494] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8efd9696-29c3-4a65-91a0-022e339016f6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.818515] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f8dc28b-64ab-4208-848a-14a48f925b3a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.850834] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17f71569-03fc-4fd7-a1ad-ea8910ef9262 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.858667] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f582e4a-4e1e-4bd7-af50-2ed58b9e430c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.872160] env[61987]: DEBUG nova.compute.provider_tree [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1057.891388] env[61987]: DEBUG nova.network.neutron [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Port 785edefc-5e84-4a34-97e1-b7c1a08f218e binding to destination host cpu-1 is already ACTIVE {{(pid=61987) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1058.000494] env[61987]: DEBUG oslo_vmware.api [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062284, 'name': ReconfigVM_Task, 'duration_secs': 5.733181} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.000840] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.001203] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Reconfigured VM to detach interface {{(pid=61987) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1058.237964] env[61987]: DEBUG nova.network.neutron [-] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.269058] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-156a2072-3d78-43b6-a11b-23f137d980bf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.278878] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db4ddc3d-3e44-4cdc-857c-e8c50ad80061 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.309523] env[61987]: DEBUG nova.compute.manager [req-49a2167e-d446-46f3-97bf-fc19a8cbf453 req-28798237-4984-4048-b6cd-b56d6d0904c9 service nova] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Detach interface failed, port_id=ca34ba16-38ec-417f-b95f-ec6e425e5f61, reason: Instance 24f8d398-beb0-4cc5-ada5-23f701adbe04 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 1058.376161] env[61987]: DEBUG nova.scheduler.client.report [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1058.741153] env[61987]: INFO nova.compute.manager [-] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Took 1.22 seconds to deallocate network for instance. [ 1058.881969] env[61987]: DEBUG oslo_concurrency.lockutils [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.232s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.882539] env[61987]: DEBUG nova.compute.manager [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1058.921300] env[61987]: DEBUG oslo_concurrency.lockutils [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "69a22e40-d469-4500-926e-0a12a233f252-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.921522] env[61987]: DEBUG oslo_concurrency.lockutils [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "69a22e40-d469-4500-926e-0a12a233f252-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.921709] env[61987]: DEBUG oslo_concurrency.lockutils [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "69a22e40-d469-4500-926e-0a12a233f252-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.248221] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.248552] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.248768] env[61987]: DEBUG nova.objects.instance [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Lazy-loading 'resources' on Instance uuid 24f8d398-beb0-4cc5-ada5-23f701adbe04 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1059.388033] env[61987]: DEBUG nova.compute.utils [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1059.389464] env[61987]: DEBUG nova.compute.manager [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1059.389646] env[61987]: DEBUG nova.network.neutron [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1059.439724] env[61987]: DEBUG nova.policy [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c5cf7a7673cc4f5aadca7960c355c394', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ea5c65ac3b143178b13731663dbd8ee', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 1059.456088] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.456964] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.456964] env[61987]: DEBUG nova.network.neutron [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1059.725904] env[61987]: DEBUG nova.network.neutron [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Successfully created port: 9f0f3f6e-5109-4c21-9c31-5636c1f60bbc {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1059.892961] env[61987]: DEBUG nova.compute.manager [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1059.918413] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14944e1c-2618-42c1-a16e-cfc12c80bae6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.933226] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f828c9-aac6-4b99-8793-84fe119945da {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.968496] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3dcdc44-ad7b-4f1e-baa3-f2555d3ae247 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.971674] env[61987]: DEBUG oslo_concurrency.lockutils [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1059.972084] env[61987]: DEBUG oslo_concurrency.lockutils [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.972084] env[61987]: DEBUG nova.network.neutron [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1059.981790] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4075a018-4817-40bf-bde8-71b444bb3ef2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.998144] env[61987]: DEBUG nova.compute.provider_tree [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1060.081583] env[61987]: DEBUG nova.compute.manager [req-ad4fb3ee-f255-4c95-8976-ce757c376966 req-9ed16e7f-55fc-4642-a72e-109c68219377 service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Received event network-changed-fe549122-49ad-4628-b106-216d787be2d4 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1060.081785] env[61987]: DEBUG nova.compute.manager [req-ad4fb3ee-f255-4c95-8976-ce757c376966 req-9ed16e7f-55fc-4642-a72e-109c68219377 service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Refreshing instance network info cache due to event network-changed-fe549122-49ad-4628-b106-216d787be2d4. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1060.081985] env[61987]: DEBUG oslo_concurrency.lockutils [req-ad4fb3ee-f255-4c95-8976-ce757c376966 req-9ed16e7f-55fc-4642-a72e-109c68219377 service nova] Acquiring lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.450071] env[61987]: INFO nova.network.neutron [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Port b67c9377-1e7d-4264-ae4c-dee82f96be5b from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1060.450541] env[61987]: DEBUG nova.network.neutron [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Updating instance_info_cache with network_info: [{"id": "fe549122-49ad-4628-b106-216d787be2d4", "address": "fa:16:3e:76:0c:4d", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe549122-49", "ovs_interfaceid": "fe549122-49ad-4628-b106-216d787be2d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.500795] env[61987]: DEBUG nova.scheduler.client.report [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1060.758620] env[61987]: DEBUG nova.network.neutron [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updating instance_info_cache with network_info: [{"id": "785edefc-5e84-4a34-97e1-b7c1a08f218e", "address": "fa:16:3e:bf:8b:d7", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap785edefc-5e", "ovs_interfaceid": "785edefc-5e84-4a34-97e1-b7c1a08f218e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.909850] env[61987]: DEBUG nova.compute.manager [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1060.935316] env[61987]: DEBUG nova.virt.hardware [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1060.935603] env[61987]: DEBUG nova.virt.hardware [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1060.935813] env[61987]: DEBUG nova.virt.hardware [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1060.936024] env[61987]: DEBUG nova.virt.hardware [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1060.936426] env[61987]: DEBUG nova.virt.hardware [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1060.936426] env[61987]: DEBUG nova.virt.hardware [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1060.936579] env[61987]: DEBUG nova.virt.hardware [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1060.936737] env[61987]: DEBUG nova.virt.hardware [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1060.936868] env[61987]: DEBUG nova.virt.hardware [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1060.937052] env[61987]: DEBUG nova.virt.hardware [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1060.937238] env[61987]: DEBUG nova.virt.hardware [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1060.938134] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a94a58a-b992-41a7-96d3-e3263822c6de {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.946969] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a29ea159-d72c-447a-a8f6-245b793cfb0b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.963229] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.965859] env[61987]: DEBUG oslo_concurrency.lockutils [req-ad4fb3ee-f255-4c95-8976-ce757c376966 req-9ed16e7f-55fc-4642-a72e-109c68219377 service nova] Acquired lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.965971] env[61987]: DEBUG nova.network.neutron [req-ad4fb3ee-f255-4c95-8976-ce757c376966 req-9ed16e7f-55fc-4642-a72e-109c68219377 service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Refreshing network info cache for port fe549122-49ad-4628-b106-216d787be2d4 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1061.006255] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.758s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.029387] env[61987]: INFO nova.scheduler.client.report [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Deleted allocations for instance 24f8d398-beb0-4cc5-ada5-23f701adbe04 [ 1061.031969] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "interface-0cba5e51-7e0a-409b-a7b2-612ab8cf5689-b67c9377-1e7d-4264-ae4c-dee82f96be5b" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.032115] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "interface-0cba5e51-7e0a-409b-a7b2-612ab8cf5689-b67c9377-1e7d-4264-ae4c-dee82f96be5b" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.032469] env[61987]: DEBUG nova.objects.instance [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lazy-loading 'flavor' on Instance uuid 0cba5e51-7e0a-409b-a7b2-612ab8cf5689 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1061.170919] env[61987]: DEBUG nova.compute.manager [req-4bd8d844-c492-4589-aa36-ba71bcc20191 req-0af56fac-15fb-46b5-826f-039ab6ed75c3 service nova] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Received event network-vif-plugged-9f0f3f6e-5109-4c21-9c31-5636c1f60bbc {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1061.171173] env[61987]: DEBUG oslo_concurrency.lockutils [req-4bd8d844-c492-4589-aa36-ba71bcc20191 req-0af56fac-15fb-46b5-826f-039ab6ed75c3 service nova] Acquiring lock "76707afe-29ae-4257-a237-ba5a47b2511b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.171388] env[61987]: DEBUG oslo_concurrency.lockutils [req-4bd8d844-c492-4589-aa36-ba71bcc20191 req-0af56fac-15fb-46b5-826f-039ab6ed75c3 service nova] Lock "76707afe-29ae-4257-a237-ba5a47b2511b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.171557] env[61987]: DEBUG oslo_concurrency.lockutils [req-4bd8d844-c492-4589-aa36-ba71bcc20191 req-0af56fac-15fb-46b5-826f-039ab6ed75c3 service nova] Lock "76707afe-29ae-4257-a237-ba5a47b2511b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.171732] env[61987]: DEBUG nova.compute.manager [req-4bd8d844-c492-4589-aa36-ba71bcc20191 req-0af56fac-15fb-46b5-826f-039ab6ed75c3 service nova] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] No waiting events found dispatching network-vif-plugged-9f0f3f6e-5109-4c21-9c31-5636c1f60bbc {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1061.171908] env[61987]: WARNING nova.compute.manager [req-4bd8d844-c492-4589-aa36-ba71bcc20191 req-0af56fac-15fb-46b5-826f-039ab6ed75c3 service nova] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Received unexpected event network-vif-plugged-9f0f3f6e-5109-4c21-9c31-5636c1f60bbc for instance with vm_state building and task_state spawning. [ 1061.261440] env[61987]: DEBUG oslo_concurrency.lockutils [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.267239] env[61987]: DEBUG nova.network.neutron [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Successfully updated port: 9f0f3f6e-5109-4c21-9c31-5636c1f60bbc {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1061.468558] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4b07bed1-9cc2-4733-8b18-53cb0a3fc142 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "interface-7be2179d-68f7-4103-a8c1-b6e5b88b8706-b67c9377-1e7d-4264-ae4c-dee82f96be5b" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.075s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.542820] env[61987]: DEBUG oslo_concurrency.lockutils [None req-1e377c54-3904-4c41-8fee-ff3f4f5fc84b tempest-ServersNegativeTestMultiTenantJSON-1330368583 tempest-ServersNegativeTestMultiTenantJSON-1330368583-project-member] Lock "24f8d398-beb0-4cc5-ada5-23f701adbe04" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.631s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.646852] env[61987]: DEBUG nova.objects.instance [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lazy-loading 'pci_requests' on Instance uuid 0cba5e51-7e0a-409b-a7b2-612ab8cf5689 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1061.764750] env[61987]: DEBUG nova.network.neutron [req-ad4fb3ee-f255-4c95-8976-ce757c376966 req-9ed16e7f-55fc-4642-a72e-109c68219377 service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Updated VIF entry in instance network info cache for port fe549122-49ad-4628-b106-216d787be2d4. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1061.765178] env[61987]: DEBUG nova.network.neutron [req-ad4fb3ee-f255-4c95-8976-ce757c376966 req-9ed16e7f-55fc-4642-a72e-109c68219377 service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Updating instance_info_cache with network_info: [{"id": "fe549122-49ad-4628-b106-216d787be2d4", "address": "fa:16:3e:76:0c:4d", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe549122-49", "ovs_interfaceid": "fe549122-49ad-4628-b106-216d787be2d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.771839] env[61987]: DEBUG oslo_concurrency.lockutils [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "refresh_cache-76707afe-29ae-4257-a237-ba5a47b2511b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.771966] env[61987]: DEBUG oslo_concurrency.lockutils [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired lock "refresh_cache-76707afe-29ae-4257-a237-ba5a47b2511b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.772123] env[61987]: DEBUG nova.network.neutron [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1061.793357] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34e9ab36-46c5-4989-a9f2-f9b39da6e585 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.812298] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0892005b-c87d-41f3-b83f-c7de53b57bbd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.819113] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updating instance '69a22e40-d469-4500-926e-0a12a233f252' progress to 83 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1062.149786] env[61987]: DEBUG nova.objects.base [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Object Instance<0cba5e51-7e0a-409b-a7b2-612ab8cf5689> lazy-loaded attributes: flavor,pci_requests {{(pid=61987) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1062.149786] env[61987]: DEBUG nova.network.neutron [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1062.267847] env[61987]: DEBUG oslo_concurrency.lockutils [req-ad4fb3ee-f255-4c95-8976-ce757c376966 req-9ed16e7f-55fc-4642-a72e-109c68219377 service nova] Releasing lock "refresh_cache-7be2179d-68f7-4103-a8c1-b6e5b88b8706" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.308209] env[61987]: DEBUG nova.compute.manager [req-f3a56d61-4073-4c45-aa34-a1106c30648e req-a39cfda9-caf2-4c66-9f02-677ed6a3c32d service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Received event network-changed-b739a840-515d-46e6-b889-5e9f8ec9ad71 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1062.308437] env[61987]: DEBUG nova.compute.manager [req-f3a56d61-4073-4c45-aa34-a1106c30648e req-a39cfda9-caf2-4c66-9f02-677ed6a3c32d service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Refreshing instance network info cache due to event network-changed-b739a840-515d-46e6-b889-5e9f8ec9ad71. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1062.308664] env[61987]: DEBUG oslo_concurrency.lockutils [req-f3a56d61-4073-4c45-aa34-a1106c30648e req-a39cfda9-caf2-4c66-9f02-677ed6a3c32d service nova] Acquiring lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.308817] env[61987]: DEBUG oslo_concurrency.lockutils [req-f3a56d61-4073-4c45-aa34-a1106c30648e req-a39cfda9-caf2-4c66-9f02-677ed6a3c32d service nova] Acquired lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.309089] env[61987]: DEBUG nova.network.neutron [req-f3a56d61-4073-4c45-aa34-a1106c30648e req-a39cfda9-caf2-4c66-9f02-677ed6a3c32d service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Refreshing network info cache for port b739a840-515d-46e6-b889-5e9f8ec9ad71 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1062.314308] env[61987]: DEBUG nova.network.neutron [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1062.325120] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1062.325388] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-85af7233-1b3c-4c18-90e6-c56f4c592ae8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.335663] env[61987]: DEBUG oslo_vmware.api [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1062.335663] env[61987]: value = "task-1062297" [ 1062.335663] env[61987]: _type = "Task" [ 1062.335663] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.343902] env[61987]: DEBUG oslo_vmware.api [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062297, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.392631] env[61987]: DEBUG nova.policy [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '697f9b1e3b3840369549a4f1ae3173ef', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd2afdc9bddba4ccab2b8fdb1f31da2b6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 1062.442094] env[61987]: DEBUG nova.network.neutron [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Updating instance_info_cache with network_info: [{"id": "9f0f3f6e-5109-4c21-9c31-5636c1f60bbc", "address": "fa:16:3e:ae:06:b5", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f0f3f6e-51", "ovs_interfaceid": "9f0f3f6e-5109-4c21-9c31-5636c1f60bbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.845587] env[61987]: DEBUG oslo_vmware.api [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062297, 'name': PowerOnVM_Task, 'duration_secs': 0.417602} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.845995] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1062.846254] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-590d2e7f-4c1f-4bfb-af2b-dcd14b6f05f4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updating instance '69a22e40-d469-4500-926e-0a12a233f252' progress to 100 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1062.944357] env[61987]: DEBUG oslo_concurrency.lockutils [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Releasing lock "refresh_cache-76707afe-29ae-4257-a237-ba5a47b2511b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.944696] env[61987]: DEBUG nova.compute.manager [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Instance network_info: |[{"id": "9f0f3f6e-5109-4c21-9c31-5636c1f60bbc", "address": "fa:16:3e:ae:06:b5", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f0f3f6e-51", "ovs_interfaceid": "9f0f3f6e-5109-4c21-9c31-5636c1f60bbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1062.947325] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:06:b5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'be8bd197-4b2b-46e7-88ea-2554b0438584', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9f0f3f6e-5109-4c21-9c31-5636c1f60bbc', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1062.952703] env[61987]: DEBUG oslo.service.loopingcall [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1062.952919] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1062.953280] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-833471c9-2a8a-4291-addc-2da2333059c0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.975494] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1062.975494] env[61987]: value = "task-1062298" [ 1062.975494] env[61987]: _type = "Task" [ 1062.975494] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.984688] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062298, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.071426] env[61987]: DEBUG nova.network.neutron [req-f3a56d61-4073-4c45-aa34-a1106c30648e req-a39cfda9-caf2-4c66-9f02-677ed6a3c32d service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Updated VIF entry in instance network info cache for port b739a840-515d-46e6-b889-5e9f8ec9ad71. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1063.071826] env[61987]: DEBUG nova.network.neutron [req-f3a56d61-4073-4c45-aa34-a1106c30648e req-a39cfda9-caf2-4c66-9f02-677ed6a3c32d service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Updating instance_info_cache with network_info: [{"id": "b739a840-515d-46e6-b889-5e9f8ec9ad71", "address": "fa:16:3e:cb:9f:9b", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb739a840-51", "ovs_interfaceid": "b739a840-515d-46e6-b889-5e9f8ec9ad71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.201532] env[61987]: DEBUG nova.compute.manager [req-b8e017bd-ff82-4438-9655-b4b4002e010e req-ce2a2ddd-a318-4e68-8966-a17e82947c25 service nova] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Received event network-changed-9f0f3f6e-5109-4c21-9c31-5636c1f60bbc {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1063.201795] env[61987]: DEBUG nova.compute.manager [req-b8e017bd-ff82-4438-9655-b4b4002e010e req-ce2a2ddd-a318-4e68-8966-a17e82947c25 service nova] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Refreshing instance network info cache due to event network-changed-9f0f3f6e-5109-4c21-9c31-5636c1f60bbc. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1063.201977] env[61987]: DEBUG oslo_concurrency.lockutils [req-b8e017bd-ff82-4438-9655-b4b4002e010e req-ce2a2ddd-a318-4e68-8966-a17e82947c25 service nova] Acquiring lock "refresh_cache-76707afe-29ae-4257-a237-ba5a47b2511b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.202148] env[61987]: DEBUG oslo_concurrency.lockutils [req-b8e017bd-ff82-4438-9655-b4b4002e010e req-ce2a2ddd-a318-4e68-8966-a17e82947c25 service nova] Acquired lock "refresh_cache-76707afe-29ae-4257-a237-ba5a47b2511b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.202460] env[61987]: DEBUG nova.network.neutron [req-b8e017bd-ff82-4438-9655-b4b4002e010e req-ce2a2ddd-a318-4e68-8966-a17e82947c25 service nova] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Refreshing network info cache for port 9f0f3f6e-5109-4c21-9c31-5636c1f60bbc {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1063.486632] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062298, 'name': CreateVM_Task, 'duration_secs': 0.278854} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.486808] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1063.488641] env[61987]: DEBUG oslo_concurrency.lockutils [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.488641] env[61987]: DEBUG oslo_concurrency.lockutils [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.488641] env[61987]: DEBUG oslo_concurrency.lockutils [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1063.488641] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-baa4cfe0-232b-4fb8-b863-2517a21917e8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.496314] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1063.496314] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52135c31-d442-6c66-244b-2108b8f79b4d" [ 1063.496314] env[61987]: _type = "Task" [ 1063.496314] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.505159] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52135c31-d442-6c66-244b-2108b8f79b4d, 'name': SearchDatastore_Task, 'duration_secs': 0.008547} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.505438] env[61987]: DEBUG oslo_concurrency.lockutils [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.505679] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1063.505973] env[61987]: DEBUG oslo_concurrency.lockutils [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.506116] env[61987]: DEBUG oslo_concurrency.lockutils [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.506484] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1063.506801] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aa24ba71-f1a8-4c8c-a1dc-591f58628091 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.514301] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1063.514514] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1063.515286] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0daba775-8c8c-4e41-9d75-c7cdac912cf4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.520500] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1063.520500] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52704d6b-6f81-90c5-38c1-8c61962836b0" [ 1063.520500] env[61987]: _type = "Task" [ 1063.520500] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.528988] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52704d6b-6f81-90c5-38c1-8c61962836b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.575137] env[61987]: DEBUG oslo_concurrency.lockutils [req-f3a56d61-4073-4c45-aa34-a1106c30648e req-a39cfda9-caf2-4c66-9f02-677ed6a3c32d service nova] Releasing lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.946125] env[61987]: DEBUG nova.network.neutron [req-b8e017bd-ff82-4438-9655-b4b4002e010e req-ce2a2ddd-a318-4e68-8966-a17e82947c25 service nova] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Updated VIF entry in instance network info cache for port 9f0f3f6e-5109-4c21-9c31-5636c1f60bbc. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1063.946521] env[61987]: DEBUG nova.network.neutron [req-b8e017bd-ff82-4438-9655-b4b4002e010e req-ce2a2ddd-a318-4e68-8966-a17e82947c25 service nova] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Updating instance_info_cache with network_info: [{"id": "9f0f3f6e-5109-4c21-9c31-5636c1f60bbc", "address": "fa:16:3e:ae:06:b5", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f0f3f6e-51", "ovs_interfaceid": "9f0f3f6e-5109-4c21-9c31-5636c1f60bbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.031634] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52704d6b-6f81-90c5-38c1-8c61962836b0, 'name': SearchDatastore_Task, 'duration_secs': 0.008521} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.032444] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-679560aa-5619-4c24-abee-f535f3ef4677 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.037847] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1064.037847] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52478576-b80c-b35e-d3dc-0526aedc3ad5" [ 1064.037847] env[61987]: _type = "Task" [ 1064.037847] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.046019] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52478576-b80c-b35e-d3dc-0526aedc3ad5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.448786] env[61987]: DEBUG oslo_concurrency.lockutils [req-b8e017bd-ff82-4438-9655-b4b4002e010e req-ce2a2ddd-a318-4e68-8966-a17e82947c25 service nova] Releasing lock "refresh_cache-76707afe-29ae-4257-a237-ba5a47b2511b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.475832] env[61987]: DEBUG nova.network.neutron [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Successfully updated port: b67c9377-1e7d-4264-ae4c-dee82f96be5b {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1064.549414] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52478576-b80c-b35e-d3dc-0526aedc3ad5, 'name': SearchDatastore_Task, 'duration_secs': 0.009327} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.549734] env[61987]: DEBUG oslo_concurrency.lockutils [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.549998] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 76707afe-29ae-4257-a237-ba5a47b2511b/76707afe-29ae-4257-a237-ba5a47b2511b.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1064.550284] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ef1ddd06-7237-4c25-a407-97deb8a6de14 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.556643] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1064.556643] env[61987]: value = "task-1062299" [ 1064.556643] env[61987]: _type = "Task" [ 1064.556643] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.564567] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062299, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.979033] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.979280] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.979884] env[61987]: DEBUG nova.network.neutron [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1065.068036] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062299, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458912} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.068398] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 76707afe-29ae-4257-a237-ba5a47b2511b/76707afe-29ae-4257-a237-ba5a47b2511b.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1065.068655] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1065.068958] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e89501c2-bb84-4666-9ba7-eb7efc44d4cb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.076182] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1065.076182] env[61987]: value = "task-1062300" [ 1065.076182] env[61987]: _type = "Task" [ 1065.076182] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.084966] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062300, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.243538] env[61987]: DEBUG nova.compute.manager [req-b945b8b8-013b-478e-bc6d-498f7af90907 req-0a9fe24d-2871-4c2b-b863-c9dd1b01c6bf service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Received event network-vif-plugged-b67c9377-1e7d-4264-ae4c-dee82f96be5b {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1065.243729] env[61987]: DEBUG oslo_concurrency.lockutils [req-b945b8b8-013b-478e-bc6d-498f7af90907 req-0a9fe24d-2871-4c2b-b863-c9dd1b01c6bf service nova] Acquiring lock "0cba5e51-7e0a-409b-a7b2-612ab8cf5689-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.244079] env[61987]: DEBUG oslo_concurrency.lockutils [req-b945b8b8-013b-478e-bc6d-498f7af90907 req-0a9fe24d-2871-4c2b-b863-c9dd1b01c6bf service nova] Lock "0cba5e51-7e0a-409b-a7b2-612ab8cf5689-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.244285] env[61987]: DEBUG oslo_concurrency.lockutils [req-b945b8b8-013b-478e-bc6d-498f7af90907 req-0a9fe24d-2871-4c2b-b863-c9dd1b01c6bf service nova] Lock "0cba5e51-7e0a-409b-a7b2-612ab8cf5689-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.244467] env[61987]: DEBUG nova.compute.manager [req-b945b8b8-013b-478e-bc6d-498f7af90907 req-0a9fe24d-2871-4c2b-b863-c9dd1b01c6bf service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] No waiting events found dispatching network-vif-plugged-b67c9377-1e7d-4264-ae4c-dee82f96be5b {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1065.244640] env[61987]: WARNING nova.compute.manager [req-b945b8b8-013b-478e-bc6d-498f7af90907 req-0a9fe24d-2871-4c2b-b863-c9dd1b01c6bf service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Received unexpected event network-vif-plugged-b67c9377-1e7d-4264-ae4c-dee82f96be5b for instance with vm_state active and task_state None. [ 1065.244813] env[61987]: DEBUG nova.compute.manager [req-b945b8b8-013b-478e-bc6d-498f7af90907 req-0a9fe24d-2871-4c2b-b863-c9dd1b01c6bf service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Received event network-changed-b67c9377-1e7d-4264-ae4c-dee82f96be5b {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1065.244979] env[61987]: DEBUG nova.compute.manager [req-b945b8b8-013b-478e-bc6d-498f7af90907 req-0a9fe24d-2871-4c2b-b863-c9dd1b01c6bf service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Refreshing instance network info cache due to event network-changed-b67c9377-1e7d-4264-ae4c-dee82f96be5b. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1065.245174] env[61987]: DEBUG oslo_concurrency.lockutils [req-b945b8b8-013b-478e-bc6d-498f7af90907 req-0a9fe24d-2871-4c2b-b863-c9dd1b01c6bf service nova] Acquiring lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.516650] env[61987]: WARNING nova.network.neutron [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] 257456d2-c9db-4303-bdfd-035bbb123901 already exists in list: networks containing: ['257456d2-c9db-4303-bdfd-035bbb123901']. ignoring it [ 1065.585297] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062300, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066352} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.585571] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1065.586367] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d475ef0-77b8-4e38-9a84-35b38c3c9bd4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.608946] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] 76707afe-29ae-4257-a237-ba5a47b2511b/76707afe-29ae-4257-a237-ba5a47b2511b.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1065.609249] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27251bd4-8498-483c-8f84-cad2b498cf51 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.630498] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1065.630498] env[61987]: value = "task-1062301" [ 1065.630498] env[61987]: _type = "Task" [ 1065.630498] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.638825] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062301, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.796196] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bb656354-268a-4d1c-bc9e-f14fd8831e7f tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "69a22e40-d469-4500-926e-0a12a233f252" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.796462] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bb656354-268a-4d1c-bc9e-f14fd8831e7f tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "69a22e40-d469-4500-926e-0a12a233f252" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.796661] env[61987]: DEBUG nova.compute.manager [None req-bb656354-268a-4d1c-bc9e-f14fd8831e7f tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Going to confirm migration 5 {{(pid=61987) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5249}} [ 1065.891271] env[61987]: DEBUG nova.network.neutron [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Updating instance_info_cache with network_info: [{"id": "b739a840-515d-46e6-b889-5e9f8ec9ad71", "address": "fa:16:3e:cb:9f:9b", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb739a840-51", "ovs_interfaceid": "b739a840-515d-46e6-b889-5e9f8ec9ad71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b67c9377-1e7d-4264-ae4c-dee82f96be5b", "address": "fa:16:3e:9e:56:e8", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb67c9377-1e", "ovs_interfaceid": "b67c9377-1e7d-4264-ae4c-dee82f96be5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.140684] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062301, 'name': ReconfigVM_Task, 'duration_secs': 0.30487} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.140988] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Reconfigured VM instance instance-00000070 to attach disk [datastore1] 76707afe-29ae-4257-a237-ba5a47b2511b/76707afe-29ae-4257-a237-ba5a47b2511b.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1066.141657] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cafd1713-410e-43f4-aadd-277846b71a6a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.149023] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1066.149023] env[61987]: value = "task-1062302" [ 1066.149023] env[61987]: _type = "Task" [ 1066.149023] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.156462] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062302, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.333018] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bb656354-268a-4d1c-bc9e-f14fd8831e7f tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.333241] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bb656354-268a-4d1c-bc9e-f14fd8831e7f tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.333427] env[61987]: DEBUG nova.network.neutron [None req-bb656354-268a-4d1c-bc9e-f14fd8831e7f tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1066.333621] env[61987]: DEBUG nova.objects.instance [None req-bb656354-268a-4d1c-bc9e-f14fd8831e7f tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lazy-loading 'info_cache' on Instance uuid 69a22e40-d469-4500-926e-0a12a233f252 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.394622] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.395301] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.395475] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.395763] env[61987]: DEBUG oslo_concurrency.lockutils [req-b945b8b8-013b-478e-bc6d-498f7af90907 req-0a9fe24d-2871-4c2b-b863-c9dd1b01c6bf service nova] Acquired lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.396024] env[61987]: DEBUG nova.network.neutron [req-b945b8b8-013b-478e-bc6d-498f7af90907 req-0a9fe24d-2871-4c2b-b863-c9dd1b01c6bf service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Refreshing network info cache for port b67c9377-1e7d-4264-ae4c-dee82f96be5b {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1066.398256] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28f26231-c306-47b3-84e4-0459942142cc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.415834] env[61987]: DEBUG nova.virt.hardware [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1066.416219] env[61987]: DEBUG nova.virt.hardware [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1066.416396] env[61987]: DEBUG nova.virt.hardware [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1066.416600] env[61987]: DEBUG nova.virt.hardware [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1066.417407] env[61987]: DEBUG nova.virt.hardware [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1066.417631] env[61987]: DEBUG nova.virt.hardware [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1066.417865] env[61987]: DEBUG nova.virt.hardware [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1066.418059] env[61987]: DEBUG nova.virt.hardware [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1066.418264] env[61987]: DEBUG nova.virt.hardware [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1066.418433] env[61987]: DEBUG nova.virt.hardware [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1066.418671] env[61987]: DEBUG nova.virt.hardware [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1066.425134] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Reconfiguring VM to attach interface {{(pid=61987) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 1066.425724] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4bae8d5e-c9d8-4695-84bd-9ca46aa3fe24 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.442462] env[61987]: DEBUG oslo_vmware.api [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1066.442462] env[61987]: value = "task-1062303" [ 1066.442462] env[61987]: _type = "Task" [ 1066.442462] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.453112] env[61987]: DEBUG oslo_vmware.api [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062303, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.661289] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062302, 'name': Rename_Task, 'duration_secs': 0.141102} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.662564] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1066.662864] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a18e2b92-3d0a-4521-a6d1-f452adfa7940 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.669892] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1066.669892] env[61987]: value = "task-1062304" [ 1066.669892] env[61987]: _type = "Task" [ 1066.669892] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.679753] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062304, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.953222] env[61987]: DEBUG oslo_vmware.api [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062303, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.185345] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062304, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.240793] env[61987]: DEBUG nova.network.neutron [req-b945b8b8-013b-478e-bc6d-498f7af90907 req-0a9fe24d-2871-4c2b-b863-c9dd1b01c6bf service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Updated VIF entry in instance network info cache for port b67c9377-1e7d-4264-ae4c-dee82f96be5b. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1067.241336] env[61987]: DEBUG nova.network.neutron [req-b945b8b8-013b-478e-bc6d-498f7af90907 req-0a9fe24d-2871-4c2b-b863-c9dd1b01c6bf service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Updating instance_info_cache with network_info: [{"id": "b739a840-515d-46e6-b889-5e9f8ec9ad71", "address": "fa:16:3e:cb:9f:9b", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb739a840-51", "ovs_interfaceid": "b739a840-515d-46e6-b889-5e9f8ec9ad71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b67c9377-1e7d-4264-ae4c-dee82f96be5b", "address": "fa:16:3e:9e:56:e8", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb67c9377-1e", "ovs_interfaceid": "b67c9377-1e7d-4264-ae4c-dee82f96be5b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.458634] env[61987]: DEBUG oslo_vmware.api [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062303, 'name': ReconfigVM_Task, 'duration_secs': 0.611839} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.459390] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.460354] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Reconfigured VM to attach interface {{(pid=61987) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 1067.681317] env[61987]: DEBUG oslo_vmware.api [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062304, 'name': PowerOnVM_Task, 'duration_secs': 0.611375} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.681639] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1067.682035] env[61987]: INFO nova.compute.manager [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Took 6.77 seconds to spawn the instance on the hypervisor. [ 1067.682035] env[61987]: DEBUG nova.compute.manager [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1067.682853] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-231d4dc1-f150-4392-9703-10e3667ccd20 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.743897] env[61987]: DEBUG oslo_concurrency.lockutils [req-b945b8b8-013b-478e-bc6d-498f7af90907 req-0a9fe24d-2871-4c2b-b863-c9dd1b01c6bf service nova] Releasing lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.965679] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d1a57161-3f0e-4ac6-a8b3-e7f7fbc86c3e tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "interface-0cba5e51-7e0a-409b-a7b2-612ab8cf5689-b67c9377-1e7d-4264-ae4c-dee82f96be5b" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.933s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.978784] env[61987]: DEBUG nova.network.neutron [None req-bb656354-268a-4d1c-bc9e-f14fd8831e7f tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updating instance_info_cache with network_info: [{"id": "785edefc-5e84-4a34-97e1-b7c1a08f218e", "address": "fa:16:3e:bf:8b:d7", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap785edefc-5e", "ovs_interfaceid": "785edefc-5e84-4a34-97e1-b7c1a08f218e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.204025] env[61987]: INFO nova.compute.manager [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Took 11.57 seconds to build instance. [ 1068.481554] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bb656354-268a-4d1c-bc9e-f14fd8831e7f tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "refresh_cache-69a22e40-d469-4500-926e-0a12a233f252" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.481554] env[61987]: DEBUG nova.objects.instance [None req-bb656354-268a-4d1c-bc9e-f14fd8831e7f tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lazy-loading 'migration_context' on Instance uuid 69a22e40-d469-4500-926e-0a12a233f252 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1068.706407] env[61987]: DEBUG oslo_concurrency.lockutils [None req-938ac6ac-07e1-4e9a-bb61-685389095c04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "76707afe-29ae-4257-a237-ba5a47b2511b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.081s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.984228] env[61987]: DEBUG nova.objects.base [None req-bb656354-268a-4d1c-bc9e-f14fd8831e7f tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Object Instance<69a22e40-d469-4500-926e-0a12a233f252> lazy-loaded attributes: info_cache,migration_context {{(pid=61987) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1068.985603] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7451b3c4-f5c0-4743-9ad8-6dc1f5c8b9ee {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.014781] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4995631a-e3d7-4abf-9c6b-bd62d1cf3a92 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.023452] env[61987]: DEBUG oslo_vmware.api [None req-bb656354-268a-4d1c-bc9e-f14fd8831e7f tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1069.023452] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5261f031-b79c-3051-2dc4-eb489166b99a" [ 1069.023452] env[61987]: _type = "Task" [ 1069.023452] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.032817] env[61987]: DEBUG oslo_vmware.api [None req-bb656354-268a-4d1c-bc9e-f14fd8831e7f tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5261f031-b79c-3051-2dc4-eb489166b99a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.175109] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "7002962c-76ed-4a8d-be2a-b2b118847fea" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.175382] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "7002962c-76ed-4a8d-be2a-b2b118847fea" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.363397] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "interface-0cba5e51-7e0a-409b-a7b2-612ab8cf5689-b67c9377-1e7d-4264-ae4c-dee82f96be5b" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.366678] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "interface-0cba5e51-7e0a-409b-a7b2-612ab8cf5689-b67c9377-1e7d-4264-ae4c-dee82f96be5b" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.002s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.533151] env[61987]: DEBUG oslo_vmware.api [None req-bb656354-268a-4d1c-bc9e-f14fd8831e7f tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5261f031-b79c-3051-2dc4-eb489166b99a, 'name': SearchDatastore_Task, 'duration_secs': 0.010053} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.533521] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bb656354-268a-4d1c-bc9e-f14fd8831e7f tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.533760] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bb656354-268a-4d1c-bc9e-f14fd8831e7f tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.678715] env[61987]: DEBUG nova.compute.utils [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1069.868867] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.869257] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.873520] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-494c4ae9-000a-434c-a023-145905d8bcc6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.898183] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8ba42f5-ec48-42f2-943a-044da7d33a1e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.928251] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Reconfiguring VM to detach interface {{(pid=61987) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 1069.929180] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c72edb7-af4a-4fcc-b661-f44547bdf2a4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.949859] env[61987]: DEBUG oslo_vmware.api [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1069.949859] env[61987]: value = "task-1062305" [ 1069.949859] env[61987]: _type = "Task" [ 1069.949859] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.958475] env[61987]: DEBUG oslo_vmware.api [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062305, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.174433] env[61987]: DEBUG nova.compute.manager [req-bb59ee3c-390d-47da-aeda-dca61801e965 req-58f8424e-5a8b-4121-90c4-fe93d4438bc9 service nova] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Received event network-changed-9f0f3f6e-5109-4c21-9c31-5636c1f60bbc {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1070.174803] env[61987]: DEBUG nova.compute.manager [req-bb59ee3c-390d-47da-aeda-dca61801e965 req-58f8424e-5a8b-4121-90c4-fe93d4438bc9 service nova] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Refreshing instance network info cache due to event network-changed-9f0f3f6e-5109-4c21-9c31-5636c1f60bbc. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1070.175268] env[61987]: DEBUG oslo_concurrency.lockutils [req-bb59ee3c-390d-47da-aeda-dca61801e965 req-58f8424e-5a8b-4121-90c4-fe93d4438bc9 service nova] Acquiring lock "refresh_cache-76707afe-29ae-4257-a237-ba5a47b2511b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.175527] env[61987]: DEBUG oslo_concurrency.lockutils [req-bb59ee3c-390d-47da-aeda-dca61801e965 req-58f8424e-5a8b-4121-90c4-fe93d4438bc9 service nova] Acquired lock "refresh_cache-76707afe-29ae-4257-a237-ba5a47b2511b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.175909] env[61987]: DEBUG nova.network.neutron [req-bb59ee3c-390d-47da-aeda-dca61801e965 req-58f8424e-5a8b-4121-90c4-fe93d4438bc9 service nova] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Refreshing network info cache for port 9f0f3f6e-5109-4c21-9c31-5636c1f60bbc {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1070.187470] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "7002962c-76ed-4a8d-be2a-b2b118847fea" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.198395] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b379056-f861-44b0-b6df-37f17829614f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.206262] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb61bb0-8806-4bad-8294-ec98ea313ddb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.239154] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3734835-59fa-4974-b732-5e89eb89fd63 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.247218] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466db7cf-d348-4445-b3c2-c5b1db06b3d9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.260733] env[61987]: DEBUG nova.compute.provider_tree [None req-bb656354-268a-4d1c-bc9e-f14fd8831e7f tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1070.460300] env[61987]: DEBUG oslo_vmware.api [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062305, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.763076] env[61987]: DEBUG nova.scheduler.client.report [None req-bb656354-268a-4d1c-bc9e-f14fd8831e7f tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1070.797553] env[61987]: DEBUG oslo_concurrency.lockutils [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "f030044b-bc6b-478c-8f63-d519aa76203c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.797797] env[61987]: DEBUG oslo_concurrency.lockutils [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "f030044b-bc6b-478c-8f63-d519aa76203c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.919309] env[61987]: DEBUG nova.network.neutron [req-bb59ee3c-390d-47da-aeda-dca61801e965 req-58f8424e-5a8b-4121-90c4-fe93d4438bc9 service nova] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Updated VIF entry in instance network info cache for port 9f0f3f6e-5109-4c21-9c31-5636c1f60bbc. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1070.919701] env[61987]: DEBUG nova.network.neutron [req-bb59ee3c-390d-47da-aeda-dca61801e965 req-58f8424e-5a8b-4121-90c4-fe93d4438bc9 service nova] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Updating instance_info_cache with network_info: [{"id": "9f0f3f6e-5109-4c21-9c31-5636c1f60bbc", "address": "fa:16:3e:ae:06:b5", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f0f3f6e-51", "ovs_interfaceid": "9f0f3f6e-5109-4c21-9c31-5636c1f60bbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.962609] env[61987]: DEBUG oslo_vmware.api [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062305, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.261066] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "7002962c-76ed-4a8d-be2a-b2b118847fea" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.261066] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "7002962c-76ed-4a8d-be2a-b2b118847fea" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.261231] env[61987]: INFO nova.compute.manager [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Attaching volume dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb to /dev/sdb [ 1071.296097] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bfa7eb6-388f-439d-8211-ba59418cdee0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.300611] env[61987]: DEBUG nova.compute.manager [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1071.306056] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec26ccf9-a046-4b65-b91c-1a26a4ebe568 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.320031] env[61987]: DEBUG nova.virt.block_device [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Updating existing volume attachment record: d8075c51-357d-4061-8781-4a3bf3c59dfa {{(pid=61987) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1071.422304] env[61987]: DEBUG oslo_concurrency.lockutils [req-bb59ee3c-390d-47da-aeda-dca61801e965 req-58f8424e-5a8b-4121-90c4-fe93d4438bc9 service nova] Releasing lock "refresh_cache-76707afe-29ae-4257-a237-ba5a47b2511b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.460601] env[61987]: DEBUG oslo_vmware.api [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062305, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.775357] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bb656354-268a-4d1c-bc9e-f14fd8831e7f tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.241s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.825650] env[61987]: DEBUG oslo_concurrency.lockutils [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.826082] env[61987]: DEBUG oslo_concurrency.lockutils [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.828195] env[61987]: INFO nova.compute.claims [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1071.961593] env[61987]: DEBUG oslo_vmware.api [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062305, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.336043] env[61987]: INFO nova.scheduler.client.report [None req-bb656354-268a-4d1c-bc9e-f14fd8831e7f tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Deleted allocation for migration c7b03f3d-2ed7-4f0f-89bf-a42b9a6f5af6 [ 1072.462124] env[61987]: DEBUG oslo_vmware.api [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062305, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.844448] env[61987]: DEBUG oslo_concurrency.lockutils [None req-bb656354-268a-4d1c-bc9e-f14fd8831e7f tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "69a22e40-d469-4500-926e-0a12a233f252" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.048s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.959377] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58a2c11d-3e40-45d8-b00a-0afab4349a96 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.969341] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1072.969637] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Cleaning up deleted instances {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11701}} [ 1072.970813] env[61987]: DEBUG oslo_vmware.api [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062305, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.974079] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55258653-3bd6-4889-9316-e0e98ad9ed76 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.005544] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de6f07b3-0d30-4705-a747-1a4f463f4a63 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.012856] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5740b68b-cb76-4014-b1a9-f02983defd43 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.025997] env[61987]: DEBUG nova.compute.provider_tree [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1073.455775] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "69a22e40-d469-4500-926e-0a12a233f252" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.456113] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "69a22e40-d469-4500-926e-0a12a233f252" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.456352] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "69a22e40-d469-4500-926e-0a12a233f252-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.456555] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "69a22e40-d469-4500-926e-0a12a233f252-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.456736] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "69a22e40-d469-4500-926e-0a12a233f252-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.462178] env[61987]: INFO nova.compute.manager [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Terminating instance [ 1073.469127] env[61987]: DEBUG oslo_vmware.api [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062305, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.485129] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] There are 54 instances to clean {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11710}} [ 1073.485307] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 24f8d398-beb0-4cc5-ada5-23f701adbe04] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1073.528746] env[61987]: DEBUG nova.scheduler.client.report [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1073.967590] env[61987]: DEBUG nova.compute.manager [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1073.967779] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1073.968076] env[61987]: DEBUG oslo_vmware.api [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062305, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.968780] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54379ccb-af89-473e-83e9-17b867b1995c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.975088] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1073.975360] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b38463a3-7e60-40a7-b51f-2b2c3b251f90 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.981333] env[61987]: DEBUG oslo_vmware.api [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1073.981333] env[61987]: value = "task-1062310" [ 1073.981333] env[61987]: _type = "Task" [ 1073.981333] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.989701] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: b172195a-79d2-4f24-b9df-320a07de965e] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1073.991532] env[61987]: DEBUG oslo_vmware.api [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062310, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.032990] env[61987]: DEBUG oslo_concurrency.lockutils [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.207s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.033551] env[61987]: DEBUG nova.compute.manager [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1074.465596] env[61987]: DEBUG oslo_vmware.api [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062305, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.489948] env[61987]: DEBUG oslo_vmware.api [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062310, 'name': PowerOffVM_Task, 'duration_secs': 0.217715} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.490231] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1074.490411] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1074.490654] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f8f4670f-b270-4a5a-bb6f-f48e84cc6e59 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.492199] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: f3773cda-41e8-4804-9319-c7e458112ddf] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1074.538686] env[61987]: DEBUG nova.compute.utils [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1074.540039] env[61987]: DEBUG nova.compute.manager [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1074.540212] env[61987]: DEBUG nova.network.neutron [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1074.553289] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1074.553505] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1074.553691] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Deleting the datastore file [datastore2] 69a22e40-d469-4500-926e-0a12a233f252 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1074.554180] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b06c5c29-0b47-456d-82e3-efd3d78b8e7b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.560382] env[61987]: DEBUG oslo_vmware.api [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1074.560382] env[61987]: value = "task-1062312" [ 1074.560382] env[61987]: _type = "Task" [ 1074.560382] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.567883] env[61987]: DEBUG oslo_vmware.api [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062312, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.578561] env[61987]: DEBUG nova.policy [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eef90ff5a24b4091ac65ceef920778c1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '90faa728219a432fa241b635071729e5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 1074.846467] env[61987]: DEBUG nova.network.neutron [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Successfully created port: 2d38bac7-bebb-462a-aac9-bbff9d711577 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1074.966907] env[61987]: DEBUG oslo_vmware.api [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062305, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.995786] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: dab6d2c5-8c04-4977-a409-b384037eabce] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1075.043277] env[61987]: DEBUG nova.compute.manager [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1075.071811] env[61987]: DEBUG oslo_vmware.api [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062312, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14037} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.072116] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1075.072326] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1075.072517] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1075.072709] env[61987]: INFO nova.compute.manager [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1075.072958] env[61987]: DEBUG oslo.service.loopingcall [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1075.073234] env[61987]: DEBUG nova.compute.manager [-] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1075.073326] env[61987]: DEBUG nova.network.neutron [-] [instance: 69a22e40-d469-4500-926e-0a12a233f252] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1075.469022] env[61987]: DEBUG oslo_vmware.api [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062305, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.498766] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 0da22c35-f11d-4498-bdb4-31bfd4c364e6] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1075.542323] env[61987]: DEBUG nova.compute.manager [req-0a65f810-dd1d-42e3-af0b-02cc74bbb5df req-1f1f5b69-32ca-4a5f-b75a-b589e76eb1c8 service nova] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Received event network-vif-deleted-785edefc-5e84-4a34-97e1-b7c1a08f218e {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1075.542540] env[61987]: INFO nova.compute.manager [req-0a65f810-dd1d-42e3-af0b-02cc74bbb5df req-1f1f5b69-32ca-4a5f-b75a-b589e76eb1c8 service nova] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Neutron deleted interface 785edefc-5e84-4a34-97e1-b7c1a08f218e; detaching it from the instance and deleting it from the info cache [ 1075.542720] env[61987]: DEBUG nova.network.neutron [req-0a65f810-dd1d-42e3-af0b-02cc74bbb5df req-1f1f5b69-32ca-4a5f-b75a-b589e76eb1c8 service nova] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.866692] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Volume attach. Driver type: vmdk {{(pid=61987) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1075.866943] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234432', 'volume_id': 'dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb', 'name': 'volume-dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7002962c-76ed-4a8d-be2a-b2b118847fea', 'attached_at': '', 'detached_at': '', 'volume_id': 'dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb', 'serial': 'dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1075.867887] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c25ee30e-9b85-4c5d-8b2e-eedcb18006be {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.883903] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc9a48c-e21c-4f87-b6f9-6639ffa1825e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.909168] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] volume-dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb/volume-dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1075.909474] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad7d1232-9578-4fae-b883-4bd88d2d3554 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.929053] env[61987]: DEBUG oslo_vmware.api [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1075.929053] env[61987]: value = "task-1062313" [ 1075.929053] env[61987]: _type = "Task" [ 1075.929053] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.936472] env[61987]: DEBUG oslo_vmware.api [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062313, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.967342] env[61987]: DEBUG oslo_vmware.api [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062305, 'name': ReconfigVM_Task, 'duration_secs': 5.755581} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.967602] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1075.967826] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Reconfigured VM to detach interface {{(pid=61987) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 1076.002142] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 54d9bbcb-d01c-4e3b-8feb-4f8942d475d4] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1076.018793] env[61987]: DEBUG nova.network.neutron [-] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.045221] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1a71a46e-2d64-4008-bd26-14b24fb9dc95 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.053342] env[61987]: DEBUG nova.compute.manager [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1076.058127] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3479cf6-f2c9-47bd-af62-06cfbc1beeb3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.079614] env[61987]: DEBUG nova.virt.hardware [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1076.079921] env[61987]: DEBUG nova.virt.hardware [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1076.080155] env[61987]: DEBUG nova.virt.hardware [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1076.080382] env[61987]: DEBUG nova.virt.hardware [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1076.080570] env[61987]: DEBUG nova.virt.hardware [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1076.080734] env[61987]: DEBUG nova.virt.hardware [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1076.080950] env[61987]: DEBUG nova.virt.hardware [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1076.081142] env[61987]: DEBUG nova.virt.hardware [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1076.081321] env[61987]: DEBUG nova.virt.hardware [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1076.081493] env[61987]: DEBUG nova.virt.hardware [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1076.081753] env[61987]: DEBUG nova.virt.hardware [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1076.082619] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca215e2-a66b-4799-81f6-868b85f1957f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.093087] env[61987]: DEBUG nova.compute.manager [req-0a65f810-dd1d-42e3-af0b-02cc74bbb5df req-1f1f5b69-32ca-4a5f-b75a-b589e76eb1c8 service nova] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Detach interface failed, port_id=785edefc-5e84-4a34-97e1-b7c1a08f218e, reason: Instance 69a22e40-d469-4500-926e-0a12a233f252 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 1076.099232] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e66179c3-8d31-4133-9238-d41d50db2470 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.248117] env[61987]: DEBUG nova.compute.manager [req-d362c67e-1722-4a20-b765-3711155e9074 req-07814f5e-b282-475a-b23a-65aff68e71ce service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Received event network-vif-plugged-2d38bac7-bebb-462a-aac9-bbff9d711577 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1076.248361] env[61987]: DEBUG oslo_concurrency.lockutils [req-d362c67e-1722-4a20-b765-3711155e9074 req-07814f5e-b282-475a-b23a-65aff68e71ce service nova] Acquiring lock "f030044b-bc6b-478c-8f63-d519aa76203c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.248576] env[61987]: DEBUG oslo_concurrency.lockutils [req-d362c67e-1722-4a20-b765-3711155e9074 req-07814f5e-b282-475a-b23a-65aff68e71ce service nova] Lock "f030044b-bc6b-478c-8f63-d519aa76203c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.248803] env[61987]: DEBUG oslo_concurrency.lockutils [req-d362c67e-1722-4a20-b765-3711155e9074 req-07814f5e-b282-475a-b23a-65aff68e71ce service nova] Lock "f030044b-bc6b-478c-8f63-d519aa76203c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.248916] env[61987]: DEBUG nova.compute.manager [req-d362c67e-1722-4a20-b765-3711155e9074 req-07814f5e-b282-475a-b23a-65aff68e71ce service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] No waiting events found dispatching network-vif-plugged-2d38bac7-bebb-462a-aac9-bbff9d711577 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1076.249352] env[61987]: WARNING nova.compute.manager [req-d362c67e-1722-4a20-b765-3711155e9074 req-07814f5e-b282-475a-b23a-65aff68e71ce service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Received unexpected event network-vif-plugged-2d38bac7-bebb-462a-aac9-bbff9d711577 for instance with vm_state building and task_state spawning. [ 1076.438327] env[61987]: DEBUG oslo_vmware.api [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062313, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.508046] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: e6236bf1-4068-4a6e-9787-cd812381e122] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1076.522061] env[61987]: INFO nova.compute.manager [-] [instance: 69a22e40-d469-4500-926e-0a12a233f252] Took 1.45 seconds to deallocate network for instance. [ 1076.797633] env[61987]: DEBUG nova.network.neutron [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Successfully updated port: 2d38bac7-bebb-462a-aac9-bbff9d711577 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1076.822972] env[61987]: DEBUG nova.compute.manager [req-2cea8e44-8f81-4b12-93c6-cb1f9220ccf9 req-ba3c5500-fa3e-4d2b-bdad-79aef7dcee06 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Received event network-changed-2d38bac7-bebb-462a-aac9-bbff9d711577 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1076.823218] env[61987]: DEBUG nova.compute.manager [req-2cea8e44-8f81-4b12-93c6-cb1f9220ccf9 req-ba3c5500-fa3e-4d2b-bdad-79aef7dcee06 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Refreshing instance network info cache due to event network-changed-2d38bac7-bebb-462a-aac9-bbff9d711577. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1076.823446] env[61987]: DEBUG oslo_concurrency.lockutils [req-2cea8e44-8f81-4b12-93c6-cb1f9220ccf9 req-ba3c5500-fa3e-4d2b-bdad-79aef7dcee06 service nova] Acquiring lock "refresh_cache-f030044b-bc6b-478c-8f63-d519aa76203c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.823819] env[61987]: DEBUG oslo_concurrency.lockutils [req-2cea8e44-8f81-4b12-93c6-cb1f9220ccf9 req-ba3c5500-fa3e-4d2b-bdad-79aef7dcee06 service nova] Acquired lock "refresh_cache-f030044b-bc6b-478c-8f63-d519aa76203c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.824096] env[61987]: DEBUG nova.network.neutron [req-2cea8e44-8f81-4b12-93c6-cb1f9220ccf9 req-ba3c5500-fa3e-4d2b-bdad-79aef7dcee06 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Refreshing network info cache for port 2d38bac7-bebb-462a-aac9-bbff9d711577 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1076.886475] env[61987]: DEBUG oslo_concurrency.lockutils [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "bf1123bf-80e6-4be3-804f-e13b906ee44d" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.886800] env[61987]: DEBUG oslo_concurrency.lockutils [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "bf1123bf-80e6-4be3-804f-e13b906ee44d" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.938768] env[61987]: DEBUG oslo_vmware.api [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062313, 'name': ReconfigVM_Task, 'duration_secs': 0.555714} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.939104] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Reconfigured VM instance instance-0000006e to attach disk [datastore1] volume-dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb/volume-dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1076.943973] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e8db8fa-e4e3-4fc4-ad09-c40e8b448852 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.959174] env[61987]: DEBUG oslo_vmware.api [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1076.959174] env[61987]: value = "task-1062314" [ 1076.959174] env[61987]: _type = "Task" [ 1076.959174] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.967378] env[61987]: DEBUG oslo_vmware.api [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062314, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.011157] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 6ec74ee9-a73b-4eac-97f2-abac3b1e3191] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1077.028578] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.028907] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.029243] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.053214] env[61987]: INFO nova.scheduler.client.report [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Deleted allocations for instance 69a22e40-d469-4500-926e-0a12a233f252 [ 1077.284336] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.284557] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquired lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.284743] env[61987]: DEBUG nova.network.neutron [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1077.300255] env[61987]: DEBUG oslo_concurrency.lockutils [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "refresh_cache-f030044b-bc6b-478c-8f63-d519aa76203c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.364807] env[61987]: DEBUG nova.network.neutron [req-2cea8e44-8f81-4b12-93c6-cb1f9220ccf9 req-ba3c5500-fa3e-4d2b-bdad-79aef7dcee06 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1077.389756] env[61987]: DEBUG nova.compute.utils [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1077.447689] env[61987]: DEBUG nova.network.neutron [req-2cea8e44-8f81-4b12-93c6-cb1f9220ccf9 req-ba3c5500-fa3e-4d2b-bdad-79aef7dcee06 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.470245] env[61987]: DEBUG oslo_vmware.api [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062314, 'name': ReconfigVM_Task, 'duration_secs': 0.145475} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.470588] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234432', 'volume_id': 'dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb', 'name': 'volume-dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7002962c-76ed-4a8d-be2a-b2b118847fea', 'attached_at': '', 'detached_at': '', 'volume_id': 'dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb', 'serial': 'dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1077.514925] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 4dc624ca-c23c-4c4f-8481-45f954e1c068] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1077.561062] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2237798a-2967-4f5a-a188-d6f2ee78a9e0 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "69a22e40-d469-4500-926e-0a12a233f252" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.104s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.773978] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "0cba5e51-7e0a-409b-a7b2-612ab8cf5689" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.774147] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "0cba5e51-7e0a-409b-a7b2-612ab8cf5689" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.774353] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "0cba5e51-7e0a-409b-a7b2-612ab8cf5689-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.774551] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "0cba5e51-7e0a-409b-a7b2-612ab8cf5689-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.774731] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "0cba5e51-7e0a-409b-a7b2-612ab8cf5689-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.776971] env[61987]: INFO nova.compute.manager [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Terminating instance [ 1077.892633] env[61987]: DEBUG oslo_concurrency.lockutils [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "bf1123bf-80e6-4be3-804f-e13b906ee44d" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.950825] env[61987]: DEBUG oslo_concurrency.lockutils [req-2cea8e44-8f81-4b12-93c6-cb1f9220ccf9 req-ba3c5500-fa3e-4d2b-bdad-79aef7dcee06 service nova] Releasing lock "refresh_cache-f030044b-bc6b-478c-8f63-d519aa76203c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.951259] env[61987]: DEBUG oslo_concurrency.lockutils [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquired lock "refresh_cache-f030044b-bc6b-478c-8f63-d519aa76203c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.951425] env[61987]: DEBUG nova.network.neutron [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1078.019498] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 70702eab-b510-4946-a426-92902b591481] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1078.095339] env[61987]: INFO nova.network.neutron [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Port b67c9377-1e7d-4264-ae4c-dee82f96be5b from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1078.095774] env[61987]: DEBUG nova.network.neutron [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Updating instance_info_cache with network_info: [{"id": "b739a840-515d-46e6-b889-5e9f8ec9ad71", "address": "fa:16:3e:cb:9f:9b", "network": {"id": "257456d2-c9db-4303-bdfd-035bbb123901", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-85273035-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2afdc9bddba4ccab2b8fdb1f31da2b6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a1c8fdf9-970c-4ae0-b6d9-f1015196b552", "external-id": "nsx-vlan-transportzone-993", "segmentation_id": 993, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb739a840-51", "ovs_interfaceid": "b739a840-515d-46e6-b889-5e9f8ec9ad71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.280909] env[61987]: DEBUG nova.compute.manager [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1078.281187] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1078.282084] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c19fcff9-bb3a-430c-8b62-a12ee5f545c7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.289914] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1078.290157] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c9e0936f-4a41-42f1-aa81-0607529609f8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.295943] env[61987]: DEBUG oslo_vmware.api [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1078.295943] env[61987]: value = "task-1062315" [ 1078.295943] env[61987]: _type = "Task" [ 1078.295943] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.303527] env[61987]: DEBUG oslo_vmware.api [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062315, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.412286] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "5912b687-2150-4133-a3a1-44ce38b3c3cb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.412537] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "5912b687-2150-4133-a3a1-44ce38b3c3cb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.495893] env[61987]: DEBUG nova.network.neutron [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1078.509441] env[61987]: DEBUG nova.objects.instance [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lazy-loading 'flavor' on Instance uuid 7002962c-76ed-4a8d-be2a-b2b118847fea {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1078.521935] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 865222ae-12da-4e5a-84af-041003b50520] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1078.598313] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Releasing lock "refresh_cache-0cba5e51-7e0a-409b-a7b2-612ab8cf5689" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.638724] env[61987]: DEBUG nova.network.neutron [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Updating instance_info_cache with network_info: [{"id": "2d38bac7-bebb-462a-aac9-bbff9d711577", "address": "fa:16:3e:f8:d4:2d", "network": {"id": "eac55241-aad5-4ddb-a01e-625c9298a8af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1411727225-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90faa728219a432fa241b635071729e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f996252-e329-42bd-a897-446dfe2b81cd", "external-id": "nsx-vlan-transportzone-535", "segmentation_id": 535, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d38bac7-be", "ovs_interfaceid": "2d38bac7-bebb-462a-aac9-bbff9d711577", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.806257] env[61987]: DEBUG oslo_vmware.api [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062315, 'name': PowerOffVM_Task, 'duration_secs': 0.211202} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.806538] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1078.806719] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1078.806965] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-741788c2-2c48-48ed-b7d3-84af5ad0a9b2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.864916] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1078.865107] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1078.865308] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Deleting the datastore file [datastore2] 0cba5e51-7e0a-409b-a7b2-612ab8cf5689 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1078.865613] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4618ec97-d20f-4a89-a44b-f423f7ba34e2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.872193] env[61987]: DEBUG oslo_vmware.api [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1078.872193] env[61987]: value = "task-1062317" [ 1078.872193] env[61987]: _type = "Task" [ 1078.872193] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.880017] env[61987]: DEBUG oslo_vmware.api [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062317, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.914931] env[61987]: DEBUG nova.compute.manager [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1078.958569] env[61987]: DEBUG oslo_concurrency.lockutils [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "bf1123bf-80e6-4be3-804f-e13b906ee44d" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.958820] env[61987]: DEBUG oslo_concurrency.lockutils [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "bf1123bf-80e6-4be3-804f-e13b906ee44d" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.959078] env[61987]: INFO nova.compute.manager [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Attaching volume 7c9a3b53-0803-45d3-9ebd-1ba92fed6944 to /dev/sdb [ 1078.994332] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd8f635-e20c-44e7-b9fd-126107c6d443 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.001080] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-175571f3-2a0e-4cdf-bf2e-d18c4a52d681 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.017969] env[61987]: DEBUG nova.virt.block_device [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Updating existing volume attachment record: fc2d5fe5-f121-4976-81ef-0a82fdbf0319 {{(pid=61987) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1079.020399] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e5de6c4f-9b7a-435c-b20d-0383a93154fb tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "7002962c-76ed-4a8d-be2a-b2b118847fea" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.759s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.025227] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: cacd3aa6-27ab-4d2b-b7ab-3381a08f3ee0] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1079.101866] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2b014b21-90e1-40d7-9253-b1d8f7576602 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "interface-0cba5e51-7e0a-409b-a7b2-612ab8cf5689-b67c9377-1e7d-4264-ae4c-dee82f96be5b" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.737s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.142351] env[61987]: DEBUG oslo_concurrency.lockutils [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Releasing lock "refresh_cache-f030044b-bc6b-478c-8f63-d519aa76203c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.142351] env[61987]: DEBUG nova.compute.manager [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Instance network_info: |[{"id": "2d38bac7-bebb-462a-aac9-bbff9d711577", "address": "fa:16:3e:f8:d4:2d", "network": {"id": "eac55241-aad5-4ddb-a01e-625c9298a8af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1411727225-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90faa728219a432fa241b635071729e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f996252-e329-42bd-a897-446dfe2b81cd", "external-id": "nsx-vlan-transportzone-535", "segmentation_id": 535, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d38bac7-be", "ovs_interfaceid": "2d38bac7-bebb-462a-aac9-bbff9d711577", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1079.142941] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f8:d4:2d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1f996252-e329-42bd-a897-446dfe2b81cd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2d38bac7-bebb-462a-aac9-bbff9d711577', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1079.150480] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Creating folder: Project (90faa728219a432fa241b635071729e5). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1079.151031] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-08f452d5-377f-4c28-b949-5297fd6c6ab1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.161852] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Created folder: Project (90faa728219a432fa241b635071729e5) in parent group-v234219. [ 1079.162064] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Creating folder: Instances. Parent ref: group-v234433. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1079.162310] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b7687529-144e-4ed8-9edc-0f4aeb81562c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.171020] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Created folder: Instances in parent group-v234433. [ 1079.171213] env[61987]: DEBUG oslo.service.loopingcall [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1079.171415] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1079.171621] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6a08b187-72b6-42be-b353-381519d4d766 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.191342] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1079.191342] env[61987]: value = "task-1062321" [ 1079.191342] env[61987]: _type = "Task" [ 1079.191342] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.203309] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062321, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.382840] env[61987]: DEBUG oslo_vmware.api [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062317, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154856} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.383012] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1079.383238] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1079.383496] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1079.383700] env[61987]: INFO nova.compute.manager [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1079.383998] env[61987]: DEBUG oslo.service.loopingcall [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1079.384290] env[61987]: DEBUG nova.compute.manager [-] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1079.384414] env[61987]: DEBUG nova.network.neutron [-] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1079.439877] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.440182] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.441880] env[61987]: INFO nova.compute.claims [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1079.529199] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: f9c82f38-7e06-4949-80a5-6eaa3c01c88c] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1079.700979] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062321, 'name': CreateVM_Task, 'duration_secs': 0.34142} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.701183] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1079.701868] env[61987]: DEBUG oslo_concurrency.lockutils [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.702066] env[61987]: DEBUG oslo_concurrency.lockutils [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.702405] env[61987]: DEBUG oslo_concurrency.lockutils [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1079.702664] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2deed275-444f-46af-9994-9f001f73e814 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.704403] env[61987]: INFO nova.compute.manager [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Rescuing [ 1079.704628] env[61987]: DEBUG oslo_concurrency.lockutils [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "refresh_cache-7002962c-76ed-4a8d-be2a-b2b118847fea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.704786] env[61987]: DEBUG oslo_concurrency.lockutils [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquired lock "refresh_cache-7002962c-76ed-4a8d-be2a-b2b118847fea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.704953] env[61987]: DEBUG nova.network.neutron [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1079.710018] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1079.710018] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52f7cc44-be84-7fed-919a-58b74d09c9e9" [ 1079.710018] env[61987]: _type = "Task" [ 1079.710018] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.718785] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52f7cc44-be84-7fed-919a-58b74d09c9e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.032821] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: a720083b-7b16-468a-a595-bfe8592e8621] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1080.078640] env[61987]: DEBUG nova.compute.manager [req-1958f407-48ff-481b-924e-46a652d05bba req-59f42bad-cc99-43ab-878e-3535db45f7b9 service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Received event network-vif-deleted-b739a840-515d-46e6-b889-5e9f8ec9ad71 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1080.078864] env[61987]: INFO nova.compute.manager [req-1958f407-48ff-481b-924e-46a652d05bba req-59f42bad-cc99-43ab-878e-3535db45f7b9 service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Neutron deleted interface b739a840-515d-46e6-b889-5e9f8ec9ad71; detaching it from the instance and deleting it from the info cache [ 1080.079067] env[61987]: DEBUG nova.network.neutron [req-1958f407-48ff-481b-924e-46a652d05bba req-59f42bad-cc99-43ab-878e-3535db45f7b9 service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.219975] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52f7cc44-be84-7fed-919a-58b74d09c9e9, 'name': SearchDatastore_Task, 'duration_secs': 0.009485} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.220320] env[61987]: DEBUG oslo_concurrency.lockutils [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.220563] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1080.220802] env[61987]: DEBUG oslo_concurrency.lockutils [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.220955] env[61987]: DEBUG oslo_concurrency.lockutils [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.221153] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1080.221672] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2154400a-d8c2-442f-9310-5a26fad608e3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.229858] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1080.230058] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1080.230771] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-030d2799-13c7-46a6-93e2-2630c56746e5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.235595] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1080.235595] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52e75449-f2a3-427a-3578-09cbe3c70e79" [ 1080.235595] env[61987]: _type = "Task" [ 1080.235595] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.245710] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52e75449-f2a3-427a-3578-09cbe3c70e79, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.468104] env[61987]: DEBUG nova.network.neutron [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Updating instance_info_cache with network_info: [{"id": "d96dede6-c439-4a14-a4d1-62c44e03ea11", "address": "fa:16:3e:d2:e0:b1", "network": {"id": "4ed4fbdc-05b0-44f8-a3d5-b7b288abd5e0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2067313775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a234a62da0e498fbe613fbcaaec3201", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd96dede6-c4", "ovs_interfaceid": "d96dede6-c439-4a14-a4d1-62c44e03ea11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.536430] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: ae1396b3-c28b-4d1c-9a3a-3dc2170847ca] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1080.553924] env[61987]: DEBUG nova.network.neutron [-] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.582496] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1efa6827-a10f-4415-aaec-b75b6cbf1bdd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.585123] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-41b07e72-5813-438b-ba43-19d17d8cdb63 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.592802] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2227dd6-187e-466d-990e-aec6b98e3698 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.598295] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2787c103-f9a8-45fb-9641-064ca71951b7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.645553] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5207fe8b-e380-4e43-8a90-3e101efd3cfc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.648680] env[61987]: DEBUG nova.compute.manager [req-1958f407-48ff-481b-924e-46a652d05bba req-59f42bad-cc99-43ab-878e-3535db45f7b9 service nova] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Detach interface failed, port_id=b739a840-515d-46e6-b889-5e9f8ec9ad71, reason: Instance 0cba5e51-7e0a-409b-a7b2-612ab8cf5689 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 1080.653940] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088c43a3-70be-454d-a6fc-e074dc230e22 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.667193] env[61987]: DEBUG nova.compute.provider_tree [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1080.746568] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52e75449-f2a3-427a-3578-09cbe3c70e79, 'name': SearchDatastore_Task, 'duration_secs': 0.008948} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.746998] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dad257b5-7f0a-4522-a560-c7f8af2793ce {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.753057] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1080.753057] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52499ffe-5b90-ef1e-1721-6ccdd8dc37dc" [ 1080.753057] env[61987]: _type = "Task" [ 1080.753057] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.759994] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52499ffe-5b90-ef1e-1721-6ccdd8dc37dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.971029] env[61987]: DEBUG oslo_concurrency.lockutils [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Releasing lock "refresh_cache-7002962c-76ed-4a8d-be2a-b2b118847fea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.038999] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: ad9041d9-1e0f-46d6-93f7-efe82197deb6] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1081.056403] env[61987]: INFO nova.compute.manager [-] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Took 1.67 seconds to deallocate network for instance. [ 1081.170576] env[61987]: DEBUG nova.scheduler.client.report [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1081.262471] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52499ffe-5b90-ef1e-1721-6ccdd8dc37dc, 'name': SearchDatastore_Task, 'duration_secs': 0.009404} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.262751] env[61987]: DEBUG oslo_concurrency.lockutils [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.263033] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] f030044b-bc6b-478c-8f63-d519aa76203c/f030044b-bc6b-478c-8f63-d519aa76203c.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1081.263293] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f8a32b9f-b6d2-4ea3-a9ac-e423cf6e67a9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.270064] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1081.270064] env[61987]: value = "task-1062323" [ 1081.270064] env[61987]: _type = "Task" [ 1081.270064] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.277779] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062323, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.542310] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 876b74ae-cae8-44a9-96ba-e2869ed43be6] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1081.562535] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1081.678088] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.237s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.678317] env[61987]: DEBUG nova.compute.manager [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1081.681595] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.119s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1081.681911] env[61987]: DEBUG nova.objects.instance [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lazy-loading 'resources' on Instance uuid 0cba5e51-7e0a-409b-a7b2-612ab8cf5689 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1081.780520] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062323, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.045699] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 286f24dd-2239-434e-8ae7-15164a0ea8b0] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1082.184912] env[61987]: DEBUG nova.compute.utils [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1082.189034] env[61987]: DEBUG nova.compute.manager [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1082.189188] env[61987]: DEBUG nova.network.neutron [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1082.231312] env[61987]: DEBUG nova.policy [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2941054f63934469bf4daa6b5a912e14', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '764082416d314c3f92eb83f576aef222', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 1082.280967] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062323, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.519229} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.281300] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] f030044b-bc6b-478c-8f63-d519aa76203c/f030044b-bc6b-478c-8f63-d519aa76203c.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1082.281467] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1082.281745] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-416ea1ce-1de2-4c28-aac6-b0549b6a052b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.287568] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1082.287568] env[61987]: value = "task-1062324" [ 1082.287568] env[61987]: _type = "Task" [ 1082.287568] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.298746] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062324, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.324725] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be24c98-f218-4072-a8fd-6b745e870fea {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.331929] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dbabc88-7ad8-4629-b009-a47363aa6c49 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.362619] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4873448a-cb83-47e2-9815-13d4b3eebb1d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.370280] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0570e876-1514-4516-af22-eb0f842a5a17 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.383213] env[61987]: DEBUG nova.compute.provider_tree [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1082.507208] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1082.507754] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0bec17d9-11ba-4305-8680-827acd8e2499 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.516696] env[61987]: DEBUG oslo_vmware.api [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1082.516696] env[61987]: value = "task-1062325" [ 1082.516696] env[61987]: _type = "Task" [ 1082.516696] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.520918] env[61987]: DEBUG nova.network.neutron [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Successfully created port: a8217ebd-b52f-4664-a191-dc7f3513a366 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1082.529768] env[61987]: DEBUG oslo_vmware.api [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062325, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.548679] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 5f45c890-b2ac-4749-a9d4-7bdc9c0bfd32] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1082.690443] env[61987]: DEBUG nova.compute.manager [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1082.796823] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062324, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066905} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.797126] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1082.797943] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f019f46-87bd-4804-9375-3ff6890b690f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.820650] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] f030044b-bc6b-478c-8f63-d519aa76203c/f030044b-bc6b-478c-8f63-d519aa76203c.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1082.820937] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a2faab21-c1b1-499b-b3de-ff7235b08985 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.840058] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1082.840058] env[61987]: value = "task-1062326" [ 1082.840058] env[61987]: _type = "Task" [ 1082.840058] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.847894] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062326, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.886631] env[61987]: DEBUG nova.scheduler.client.report [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1083.028261] env[61987]: DEBUG oslo_vmware.api [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062325, 'name': PowerOffVM_Task, 'duration_secs': 0.230231} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.028261] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1083.028753] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cda5970-379b-4d6d-8a25-92083aa0e957 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.051048] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fa0004b-397a-4dba-89ab-a9268446ed05 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.053584] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 1b08af4a-8e3e-405e-8526-1e0a37b1f0cd] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1083.085015] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1083.085552] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ccffea4-21aa-4908-a5e7-9d44288684ca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.091727] env[61987]: DEBUG oslo_vmware.api [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1083.091727] env[61987]: value = "task-1062327" [ 1083.091727] env[61987]: _type = "Task" [ 1083.091727] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.102944] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] VM already powered off {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1083.103172] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1083.103425] env[61987]: DEBUG oslo_concurrency.lockutils [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.103583] env[61987]: DEBUG oslo_concurrency.lockutils [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.103763] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1083.104006] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-06bfffa5-6916-42f2-8e92-12bc79441a72 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.116640] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1083.116824] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1083.117596] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbef8688-016e-4ff6-9c8a-b1818d4e21fb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.122430] env[61987]: DEBUG oslo_vmware.api [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1083.122430] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52100868-8d23-9aaa-9851-bb6667acdf6a" [ 1083.122430] env[61987]: _type = "Task" [ 1083.122430] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.129931] env[61987]: DEBUG oslo_vmware.api [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52100868-8d23-9aaa-9851-bb6667acdf6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.351973] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062326, 'name': ReconfigVM_Task, 'duration_secs': 0.260806} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.352866] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Reconfigured VM instance instance-00000071 to attach disk [datastore2] f030044b-bc6b-478c-8f63-d519aa76203c/f030044b-bc6b-478c-8f63-d519aa76203c.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1083.353353] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6d581251-5ed6-4ce2-8657-ff1d01f30453 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.359661] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1083.359661] env[61987]: value = "task-1062328" [ 1083.359661] env[61987]: _type = "Task" [ 1083.359661] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.368307] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062328, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.391358] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.710s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.491271] env[61987]: INFO nova.scheduler.client.report [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Deleted allocations for instance 0cba5e51-7e0a-409b-a7b2-612ab8cf5689 [ 1083.557187] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: efdf62cc-9a3f-4b6b-a119-29a8096848ee] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1083.566154] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Volume attach. Driver type: vmdk {{(pid=61987) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1083.566489] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234435', 'volume_id': '7c9a3b53-0803-45d3-9ebd-1ba92fed6944', 'name': 'volume-7c9a3b53-0803-45d3-9ebd-1ba92fed6944', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bf1123bf-80e6-4be3-804f-e13b906ee44d', 'attached_at': '', 'detached_at': '', 'volume_id': '7c9a3b53-0803-45d3-9ebd-1ba92fed6944', 'serial': '7c9a3b53-0803-45d3-9ebd-1ba92fed6944'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1083.567361] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c9b9b1f-d3ab-444f-884c-bbed88b00d5e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.587443] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0ac03a4-9a4b-467e-bbbd-007a9e52d7b2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.613237] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] volume-7c9a3b53-0803-45d3-9ebd-1ba92fed6944/volume-7c9a3b53-0803-45d3-9ebd-1ba92fed6944.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1083.613762] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32e93a73-9258-4560-a075-6014fb4d0381 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.639936] env[61987]: DEBUG oslo_vmware.api [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52100868-8d23-9aaa-9851-bb6667acdf6a, 'name': SearchDatastore_Task, 'duration_secs': 0.014536} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.641719] env[61987]: DEBUG oslo_vmware.api [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 1083.641719] env[61987]: value = "task-1062329" [ 1083.641719] env[61987]: _type = "Task" [ 1083.641719] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.641993] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2051caf7-cb84-4cdf-a72c-ff3dfc9b06dd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.650902] env[61987]: DEBUG oslo_vmware.api [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1083.650902] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]524de0ef-560d-c9a6-aa41-be00abc2e284" [ 1083.650902] env[61987]: _type = "Task" [ 1083.650902] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.654412] env[61987]: DEBUG oslo_vmware.api [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062329, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.663077] env[61987]: DEBUG oslo_vmware.api [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524de0ef-560d-c9a6-aa41-be00abc2e284, 'name': SearchDatastore_Task, 'duration_secs': 0.010154} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.663334] env[61987]: DEBUG oslo_concurrency.lockutils [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.663604] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 7002962c-76ed-4a8d-be2a-b2b118847fea/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk. {{(pid=61987) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1083.663858] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a080b86b-31d9-46d6-8427-6db252ac303f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.670317] env[61987]: DEBUG oslo_vmware.api [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1083.670317] env[61987]: value = "task-1062330" [ 1083.670317] env[61987]: _type = "Task" [ 1083.670317] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.678044] env[61987]: DEBUG oslo_vmware.api [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062330, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.700925] env[61987]: DEBUG nova.compute.manager [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1083.722561] env[61987]: DEBUG nova.virt.hardware [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1083.722561] env[61987]: DEBUG nova.virt.hardware [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1083.722561] env[61987]: DEBUG nova.virt.hardware [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1083.722904] env[61987]: DEBUG nova.virt.hardware [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1083.722904] env[61987]: DEBUG nova.virt.hardware [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1083.723015] env[61987]: DEBUG nova.virt.hardware [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1083.723195] env[61987]: DEBUG nova.virt.hardware [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1083.723367] env[61987]: DEBUG nova.virt.hardware [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1083.723546] env[61987]: DEBUG nova.virt.hardware [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1083.723718] env[61987]: DEBUG nova.virt.hardware [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1083.723899] env[61987]: DEBUG nova.virt.hardware [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1083.724810] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b4d36df-3487-47e5-9db8-94c3e1b5a213 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.732467] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ee8bbc-c8a4-4f4c-bdfa-a8c761fa8b62 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.870429] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062328, 'name': Rename_Task, 'duration_secs': 0.161931} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.870717] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1083.870992] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-74b5d565-ed22-4062-ac60-ec32dffcfa84 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.877254] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1083.877254] env[61987]: value = "task-1062331" [ 1083.877254] env[61987]: _type = "Task" [ 1083.877254] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.886391] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062331, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.000437] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d26e3c0e-9974-4c42-94c4-cddab193b1dc tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "0cba5e51-7e0a-409b-a7b2-612ab8cf5689" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.226s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.061211] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: d11104e7-4a5c-44c3-bd92-d36c587da794] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1084.087211] env[61987]: DEBUG nova.compute.manager [req-7a4bba59-162c-4356-948a-3895955838a8 req-25b21d92-f00c-4512-a036-04866bb865e8 service nova] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Received event network-vif-plugged-a8217ebd-b52f-4664-a191-dc7f3513a366 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1084.087814] env[61987]: DEBUG oslo_concurrency.lockutils [req-7a4bba59-162c-4356-948a-3895955838a8 req-25b21d92-f00c-4512-a036-04866bb865e8 service nova] Acquiring lock "5912b687-2150-4133-a3a1-44ce38b3c3cb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.088086] env[61987]: DEBUG oslo_concurrency.lockutils [req-7a4bba59-162c-4356-948a-3895955838a8 req-25b21d92-f00c-4512-a036-04866bb865e8 service nova] Lock "5912b687-2150-4133-a3a1-44ce38b3c3cb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.088302] env[61987]: DEBUG oslo_concurrency.lockutils [req-7a4bba59-162c-4356-948a-3895955838a8 req-25b21d92-f00c-4512-a036-04866bb865e8 service nova] Lock "5912b687-2150-4133-a3a1-44ce38b3c3cb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.088495] env[61987]: DEBUG nova.compute.manager [req-7a4bba59-162c-4356-948a-3895955838a8 req-25b21d92-f00c-4512-a036-04866bb865e8 service nova] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] No waiting events found dispatching network-vif-plugged-a8217ebd-b52f-4664-a191-dc7f3513a366 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1084.088679] env[61987]: WARNING nova.compute.manager [req-7a4bba59-162c-4356-948a-3895955838a8 req-25b21d92-f00c-4512-a036-04866bb865e8 service nova] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Received unexpected event network-vif-plugged-a8217ebd-b52f-4664-a191-dc7f3513a366 for instance with vm_state building and task_state spawning. [ 1084.154102] env[61987]: DEBUG oslo_vmware.api [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062329, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.179934] env[61987]: DEBUG oslo_vmware.api [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062330, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448762} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.180248] env[61987]: INFO nova.virt.vmwareapi.ds_util [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 7002962c-76ed-4a8d-be2a-b2b118847fea/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk. [ 1084.181026] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48790ae4-9c5d-438a-9160-ebc852c07f2e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.213322] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] 7002962c-76ed-4a8d-be2a-b2b118847fea/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1084.213646] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a44e41ac-e32c-4696-b0a7-d2746739c62b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.233674] env[61987]: DEBUG oslo_vmware.api [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1084.233674] env[61987]: value = "task-1062332" [ 1084.233674] env[61987]: _type = "Task" [ 1084.233674] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.241700] env[61987]: DEBUG oslo_vmware.api [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062332, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.246649] env[61987]: DEBUG nova.network.neutron [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Successfully updated port: a8217ebd-b52f-4664-a191-dc7f3513a366 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1084.388348] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062331, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.518211] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "7be2179d-68f7-4103-a8c1-b6e5b88b8706" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.518506] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "7be2179d-68f7-4103-a8c1-b6e5b88b8706" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.518814] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "7be2179d-68f7-4103-a8c1-b6e5b88b8706-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.519095] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "7be2179d-68f7-4103-a8c1-b6e5b88b8706-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.519329] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "7be2179d-68f7-4103-a8c1-b6e5b88b8706-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.521722] env[61987]: INFO nova.compute.manager [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Terminating instance [ 1084.564809] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: d8b9e172-c3bf-47d9-969b-7f33475d2d60] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1084.656617] env[61987]: DEBUG oslo_vmware.api [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062329, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.743807] env[61987]: DEBUG oslo_vmware.api [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062332, 'name': ReconfigVM_Task, 'duration_secs': 0.414254} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.744208] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Reconfigured VM instance instance-0000006e to attach disk [datastore1] 7002962c-76ed-4a8d-be2a-b2b118847fea/7bf8969f-d179-42af-9e15-527d5449f929-rescue.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1084.745247] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f16f5ff8-e4dc-45f6-b0e4-95c1001bd28b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.748772] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "refresh_cache-5912b687-2150-4133-a3a1-44ce38b3c3cb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.748913] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "refresh_cache-5912b687-2150-4133-a3a1-44ce38b3c3cb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.749072] env[61987]: DEBUG nova.network.neutron [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1084.776532] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3851ed32-73cb-40e2-bbc1-d5cd5a1668d8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.792744] env[61987]: DEBUG oslo_vmware.api [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1084.792744] env[61987]: value = "task-1062333" [ 1084.792744] env[61987]: _type = "Task" [ 1084.792744] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.800772] env[61987]: DEBUG oslo_vmware.api [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062333, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.889931] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062331, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.027070] env[61987]: DEBUG nova.compute.manager [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1085.027070] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1085.028034] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8877d9-4077-4b55-b3fd-0184e273f16d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.035820] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1085.036102] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5cbd57a0-1a15-4d40-80d6-c93aaf759457 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.043235] env[61987]: DEBUG oslo_vmware.api [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1085.043235] env[61987]: value = "task-1062334" [ 1085.043235] env[61987]: _type = "Task" [ 1085.043235] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.051319] env[61987]: DEBUG oslo_vmware.api [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062334, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.068082] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 0422be0a-213b-41e9-b850-c823b52a90ca] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1085.156272] env[61987]: DEBUG oslo_vmware.api [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062329, 'name': ReconfigVM_Task, 'duration_secs': 1.421133} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.156894] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Reconfigured VM instance instance-0000006d to attach disk [datastore2] volume-7c9a3b53-0803-45d3-9ebd-1ba92fed6944/volume-7c9a3b53-0803-45d3-9ebd-1ba92fed6944.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1085.161541] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-695c8582-65b1-486a-a038-a158b4dd9c7d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.176837] env[61987]: DEBUG oslo_vmware.api [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 1085.176837] env[61987]: value = "task-1062335" [ 1085.176837] env[61987]: _type = "Task" [ 1085.176837] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.184522] env[61987]: DEBUG oslo_vmware.api [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062335, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.302955] env[61987]: DEBUG oslo_vmware.api [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062333, 'name': ReconfigVM_Task, 'duration_secs': 0.156074} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.303745] env[61987]: DEBUG nova.network.neutron [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1085.305966] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1085.306026] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-837a249c-9657-42e8-bae9-9a037dd959fe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.312790] env[61987]: DEBUG oslo_vmware.api [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1085.312790] env[61987]: value = "task-1062336" [ 1085.312790] env[61987]: _type = "Task" [ 1085.312790] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.321187] env[61987]: DEBUG oslo_vmware.api [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062336, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.387793] env[61987]: DEBUG oslo_vmware.api [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062331, 'name': PowerOnVM_Task, 'duration_secs': 1.196545} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.388315] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1085.388748] env[61987]: INFO nova.compute.manager [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Took 9.33 seconds to spawn the instance on the hypervisor. [ 1085.388835] env[61987]: DEBUG nova.compute.manager [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1085.389586] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a84c056e-1838-4dcf-9327-376ac8980c91 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.468119] env[61987]: DEBUG nova.network.neutron [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Updating instance_info_cache with network_info: [{"id": "a8217ebd-b52f-4664-a191-dc7f3513a366", "address": "fa:16:3e:49:1b:98", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8217ebd-b5", "ovs_interfaceid": "a8217ebd-b52f-4664-a191-dc7f3513a366", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.553855] env[61987]: DEBUG oslo_vmware.api [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062334, 'name': PowerOffVM_Task, 'duration_secs': 0.321259} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.554159] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1085.554342] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1085.554649] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-255660ea-c8e1-48bf-9d79-0016dc8df062 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.571828] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 212e4860-751e-43f7-80c6-25b6784e3541] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1085.614088] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1085.614364] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1085.614556] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Deleting the datastore file [datastore1] 7be2179d-68f7-4103-a8c1-b6e5b88b8706 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1085.614823] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8759dcb2-a3a7-444b-943f-c6e99a3fba1a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.621133] env[61987]: DEBUG oslo_vmware.api [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for the task: (returnval){ [ 1085.621133] env[61987]: value = "task-1062338" [ 1085.621133] env[61987]: _type = "Task" [ 1085.621133] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.630515] env[61987]: DEBUG oslo_vmware.api [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062338, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.686735] env[61987]: DEBUG oslo_vmware.api [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062335, 'name': ReconfigVM_Task, 'duration_secs': 0.195463} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.687068] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234435', 'volume_id': '7c9a3b53-0803-45d3-9ebd-1ba92fed6944', 'name': 'volume-7c9a3b53-0803-45d3-9ebd-1ba92fed6944', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bf1123bf-80e6-4be3-804f-e13b906ee44d', 'attached_at': '', 'detached_at': '', 'volume_id': '7c9a3b53-0803-45d3-9ebd-1ba92fed6944', 'serial': '7c9a3b53-0803-45d3-9ebd-1ba92fed6944'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1085.823492] env[61987]: DEBUG oslo_vmware.api [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062336, 'name': PowerOnVM_Task, 'duration_secs': 0.479242} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.825410] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1085.826703] env[61987]: DEBUG nova.compute.manager [None req-33dda7bb-3d91-4fca-b675-d95e77c33d21 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1085.827497] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a937bad6-e51f-4e1f-8131-a5849d6a390a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.909811] env[61987]: INFO nova.compute.manager [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Took 14.10 seconds to build instance. [ 1085.970560] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "refresh_cache-5912b687-2150-4133-a3a1-44ce38b3c3cb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.970953] env[61987]: DEBUG nova.compute.manager [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Instance network_info: |[{"id": "a8217ebd-b52f-4664-a191-dc7f3513a366", "address": "fa:16:3e:49:1b:98", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8217ebd-b5", "ovs_interfaceid": "a8217ebd-b52f-4664-a191-dc7f3513a366", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1085.971419] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:49:1b:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2a111da1-447a-4722-9662-b0d6a2886a65', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a8217ebd-b52f-4664-a191-dc7f3513a366', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1085.979629] env[61987]: DEBUG oslo.service.loopingcall [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1085.979841] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1085.980186] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a726467d-1f98-4fb3-a33b-ef2d4f6d1100 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.001601] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1086.001601] env[61987]: value = "task-1062339" [ 1086.001601] env[61987]: _type = "Task" [ 1086.001601] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.009214] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062339, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.075631] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 769ecb4d-8fdd-4993-8a29-84c30d1089b2] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1086.113926] env[61987]: DEBUG nova.compute.manager [req-8bdf5fef-56b7-45c9-be89-c13be6ab8864 req-90a4c828-562d-4116-acb9-aeeb49d99429 service nova] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Received event network-changed-a8217ebd-b52f-4664-a191-dc7f3513a366 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1086.115019] env[61987]: DEBUG nova.compute.manager [req-8bdf5fef-56b7-45c9-be89-c13be6ab8864 req-90a4c828-562d-4116-acb9-aeeb49d99429 service nova] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Refreshing instance network info cache due to event network-changed-a8217ebd-b52f-4664-a191-dc7f3513a366. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1086.115019] env[61987]: DEBUG oslo_concurrency.lockutils [req-8bdf5fef-56b7-45c9-be89-c13be6ab8864 req-90a4c828-562d-4116-acb9-aeeb49d99429 service nova] Acquiring lock "refresh_cache-5912b687-2150-4133-a3a1-44ce38b3c3cb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.115019] env[61987]: DEBUG oslo_concurrency.lockutils [req-8bdf5fef-56b7-45c9-be89-c13be6ab8864 req-90a4c828-562d-4116-acb9-aeeb49d99429 service nova] Acquired lock "refresh_cache-5912b687-2150-4133-a3a1-44ce38b3c3cb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.115019] env[61987]: DEBUG nova.network.neutron [req-8bdf5fef-56b7-45c9-be89-c13be6ab8864 req-90a4c828-562d-4116-acb9-aeeb49d99429 service nova] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Refreshing network info cache for port a8217ebd-b52f-4664-a191-dc7f3513a366 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1086.133654] env[61987]: DEBUG oslo_vmware.api [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Task: {'id': task-1062338, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169152} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.133654] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1086.133654] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1086.133654] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1086.133654] env[61987]: INFO nova.compute.manager [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1086.133654] env[61987]: DEBUG oslo.service.loopingcall [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1086.133654] env[61987]: DEBUG nova.compute.manager [-] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1086.133654] env[61987]: DEBUG nova.network.neutron [-] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1086.412160] env[61987]: DEBUG oslo_concurrency.lockutils [None req-823d6e75-0916-4f80-992a-d6ecf0f260f0 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "f030044b-bc6b-478c-8f63-d519aa76203c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.614s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.509165] env[61987]: DEBUG nova.compute.manager [req-39349a6a-3341-4c02-961c-f7cf18e81c41 req-373ef70f-26d8-480f-9a68-354dea4880d3 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Received event network-changed-2d38bac7-bebb-462a-aac9-bbff9d711577 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1086.509223] env[61987]: DEBUG nova.compute.manager [req-39349a6a-3341-4c02-961c-f7cf18e81c41 req-373ef70f-26d8-480f-9a68-354dea4880d3 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Refreshing instance network info cache due to event network-changed-2d38bac7-bebb-462a-aac9-bbff9d711577. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1086.509465] env[61987]: DEBUG oslo_concurrency.lockutils [req-39349a6a-3341-4c02-961c-f7cf18e81c41 req-373ef70f-26d8-480f-9a68-354dea4880d3 service nova] Acquiring lock "refresh_cache-f030044b-bc6b-478c-8f63-d519aa76203c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.509573] env[61987]: DEBUG oslo_concurrency.lockutils [req-39349a6a-3341-4c02-961c-f7cf18e81c41 req-373ef70f-26d8-480f-9a68-354dea4880d3 service nova] Acquired lock "refresh_cache-f030044b-bc6b-478c-8f63-d519aa76203c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.509743] env[61987]: DEBUG nova.network.neutron [req-39349a6a-3341-4c02-961c-f7cf18e81c41 req-373ef70f-26d8-480f-9a68-354dea4880d3 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Refreshing network info cache for port 2d38bac7-bebb-462a-aac9-bbff9d711577 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1086.514865] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062339, 'name': CreateVM_Task, 'duration_secs': 0.322793} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.515221] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1086.515894] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.516089] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.516463] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1086.516922] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9223eb67-bfdb-4010-b5d8-6c20e4ab6b4a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.522092] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1086.522092] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52b26cb6-fd44-878f-7081-713fa55878a9" [ 1086.522092] env[61987]: _type = "Task" [ 1086.522092] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.531142] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b26cb6-fd44-878f-7081-713fa55878a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.579036] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 455562c7-9fc1-4625-95b0-5eb8c9c8ccaa] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1086.733613] env[61987]: DEBUG nova.objects.instance [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lazy-loading 'flavor' on Instance uuid bf1123bf-80e6-4be3-804f-e13b906ee44d {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1086.827832] env[61987]: DEBUG nova.network.neutron [req-8bdf5fef-56b7-45c9-be89-c13be6ab8864 req-90a4c828-562d-4116-acb9-aeeb49d99429 service nova] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Updated VIF entry in instance network info cache for port a8217ebd-b52f-4664-a191-dc7f3513a366. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1086.828276] env[61987]: DEBUG nova.network.neutron [req-8bdf5fef-56b7-45c9-be89-c13be6ab8864 req-90a4c828-562d-4116-acb9-aeeb49d99429 service nova] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Updating instance_info_cache with network_info: [{"id": "a8217ebd-b52f-4664-a191-dc7f3513a366", "address": "fa:16:3e:49:1b:98", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8217ebd-b5", "ovs_interfaceid": "a8217ebd-b52f-4664-a191-dc7f3513a366", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.872851] env[61987]: DEBUG nova.network.neutron [-] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.986860] env[61987]: INFO nova.compute.manager [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Unrescuing [ 1086.987245] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "refresh_cache-7002962c-76ed-4a8d-be2a-b2b118847fea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1086.987512] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquired lock "refresh_cache-7002962c-76ed-4a8d-be2a-b2b118847fea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1086.987745] env[61987]: DEBUG nova.network.neutron [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1087.033106] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b26cb6-fd44-878f-7081-713fa55878a9, 'name': SearchDatastore_Task, 'duration_secs': 0.009944} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.033434] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.033678] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1087.033916] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.034081] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.034274] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1087.034527] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fd77b8c9-e27e-44fb-b5d0-2c6f6ccc62c1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.042139] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1087.042328] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1087.043034] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb0d5b40-e960-418c-80e5-77a8928fe869 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.049975] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1087.049975] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]521b36c8-e63d-ea24-2f7a-182a8585c34c" [ 1087.049975] env[61987]: _type = "Task" [ 1087.049975] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.057911] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]521b36c8-e63d-ea24-2f7a-182a8585c34c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.083671] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 1bba4036-5211-4a43-b2e6-5276851335e2] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1087.217271] env[61987]: DEBUG nova.network.neutron [req-39349a6a-3341-4c02-961c-f7cf18e81c41 req-373ef70f-26d8-480f-9a68-354dea4880d3 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Updated VIF entry in instance network info cache for port 2d38bac7-bebb-462a-aac9-bbff9d711577. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1087.217837] env[61987]: DEBUG nova.network.neutron [req-39349a6a-3341-4c02-961c-f7cf18e81c41 req-373ef70f-26d8-480f-9a68-354dea4880d3 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Updating instance_info_cache with network_info: [{"id": "2d38bac7-bebb-462a-aac9-bbff9d711577", "address": "fa:16:3e:f8:d4:2d", "network": {"id": "eac55241-aad5-4ddb-a01e-625c9298a8af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1411727225-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90faa728219a432fa241b635071729e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f996252-e329-42bd-a897-446dfe2b81cd", "external-id": "nsx-vlan-transportzone-535", "segmentation_id": 535, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d38bac7-be", "ovs_interfaceid": "2d38bac7-bebb-462a-aac9-bbff9d711577", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.240362] env[61987]: DEBUG oslo_concurrency.lockutils [None req-111df5ff-73aa-42ef-b798-ace5eb553302 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "bf1123bf-80e6-4be3-804f-e13b906ee44d" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.281s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.331344] env[61987]: DEBUG oslo_concurrency.lockutils [req-8bdf5fef-56b7-45c9-be89-c13be6ab8864 req-90a4c828-562d-4116-acb9-aeeb49d99429 service nova] Releasing lock "refresh_cache-5912b687-2150-4133-a3a1-44ce38b3c3cb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.378240] env[61987]: INFO nova.compute.manager [-] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Took 1.24 seconds to deallocate network for instance. [ 1087.459068] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b9c53053-20de-4495-8a51-051f14859ec1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "bf1123bf-80e6-4be3-804f-e13b906ee44d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.459068] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b9c53053-20de-4495-8a51-051f14859ec1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "bf1123bf-80e6-4be3-804f-e13b906ee44d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.559776] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]521b36c8-e63d-ea24-2f7a-182a8585c34c, 'name': SearchDatastore_Task, 'duration_secs': 0.007874} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.562776] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-856fa5ca-9b13-4c1f-b1f9-3e4c71948835 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.568144] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1087.568144] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]528f809b-c2fa-db53-dee3-23273055264b" [ 1087.568144] env[61987]: _type = "Task" [ 1087.568144] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.576504] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]528f809b-c2fa-db53-dee3-23273055264b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.587075] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: ea8bb2f5-06d7-4fbe-bf21-9dd2d1fd3e05] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1087.716173] env[61987]: DEBUG nova.network.neutron [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Updating instance_info_cache with network_info: [{"id": "d96dede6-c439-4a14-a4d1-62c44e03ea11", "address": "fa:16:3e:d2:e0:b1", "network": {"id": "4ed4fbdc-05b0-44f8-a3d5-b7b288abd5e0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2067313775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a234a62da0e498fbe613fbcaaec3201", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd96dede6-c4", "ovs_interfaceid": "d96dede6-c439-4a14-a4d1-62c44e03ea11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.720880] env[61987]: DEBUG oslo_concurrency.lockutils [req-39349a6a-3341-4c02-961c-f7cf18e81c41 req-373ef70f-26d8-480f-9a68-354dea4880d3 service nova] Releasing lock "refresh_cache-f030044b-bc6b-478c-8f63-d519aa76203c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1087.887251] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.887566] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.887787] env[61987]: DEBUG nova.objects.instance [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lazy-loading 'resources' on Instance uuid 7be2179d-68f7-4103-a8c1-b6e5b88b8706 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.961181] env[61987]: INFO nova.compute.manager [None req-b9c53053-20de-4495-8a51-051f14859ec1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Detaching volume 7c9a3b53-0803-45d3-9ebd-1ba92fed6944 [ 1087.998856] env[61987]: INFO nova.virt.block_device [None req-b9c53053-20de-4495-8a51-051f14859ec1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Attempting to driver detach volume 7c9a3b53-0803-45d3-9ebd-1ba92fed6944 from mountpoint /dev/sdb [ 1087.998856] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9c53053-20de-4495-8a51-051f14859ec1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Volume detach. Driver type: vmdk {{(pid=61987) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1087.999017] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9c53053-20de-4495-8a51-051f14859ec1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234435', 'volume_id': '7c9a3b53-0803-45d3-9ebd-1ba92fed6944', 'name': 'volume-7c9a3b53-0803-45d3-9ebd-1ba92fed6944', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bf1123bf-80e6-4be3-804f-e13b906ee44d', 'attached_at': '', 'detached_at': '', 'volume_id': '7c9a3b53-0803-45d3-9ebd-1ba92fed6944', 'serial': '7c9a3b53-0803-45d3-9ebd-1ba92fed6944'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1087.999914] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a3bdb6-9d0f-4add-a952-729a165a6807 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.021985] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1749d81-4aaf-4303-b4db-913af0c41496 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.030189] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-623101b1-64de-46c3-ba4e-957b5c76041c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.050175] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61bbe8ed-6d33-4eac-9d26-a8347d634881 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.065254] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9c53053-20de-4495-8a51-051f14859ec1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] The volume has not been displaced from its original location: [datastore2] volume-7c9a3b53-0803-45d3-9ebd-1ba92fed6944/volume-7c9a3b53-0803-45d3-9ebd-1ba92fed6944.vmdk. No consolidation needed. {{(pid=61987) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1088.070375] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9c53053-20de-4495-8a51-051f14859ec1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Reconfiguring VM instance instance-0000006d to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1088.070764] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b634b216-f12d-4532-b986-ff67569821a2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.092614] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 742297eb-c45d-4f45-85d2-e9bdd3106d1e] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1088.094504] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]528f809b-c2fa-db53-dee3-23273055264b, 'name': SearchDatastore_Task, 'duration_secs': 0.021876} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.095973] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.096257] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 5912b687-2150-4133-a3a1-44ce38b3c3cb/5912b687-2150-4133-a3a1-44ce38b3c3cb.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1088.096791] env[61987]: DEBUG oslo_vmware.api [None req-b9c53053-20de-4495-8a51-051f14859ec1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 1088.096791] env[61987]: value = "task-1062340" [ 1088.096791] env[61987]: _type = "Task" [ 1088.096791] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.096791] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-03e91ad2-aa7b-45d1-8414-8b544b8edba4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.107164] env[61987]: DEBUG oslo_vmware.api [None req-b9c53053-20de-4495-8a51-051f14859ec1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062340, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.108394] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1088.108394] env[61987]: value = "task-1062341" [ 1088.108394] env[61987]: _type = "Task" [ 1088.108394] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.116077] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062341, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.142173] env[61987]: DEBUG nova.compute.manager [req-5523c8ca-fb08-458d-b744-3317f39ff91e req-82860bdb-d10d-4766-bf29-f11f03d18d04 service nova] [instance: 7be2179d-68f7-4103-a8c1-b6e5b88b8706] Received event network-vif-deleted-fe549122-49ad-4628-b106-216d787be2d4 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1088.218775] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Releasing lock "refresh_cache-7002962c-76ed-4a8d-be2a-b2b118847fea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.219507] env[61987]: DEBUG nova.objects.instance [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lazy-loading 'flavor' on Instance uuid 7002962c-76ed-4a8d-be2a-b2b118847fea {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1088.505626] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2bb9ef-c045-43eb-bebb-5ec465b84e7c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.513655] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5768973-51a2-4c38-bdf9-82f313bf354d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.545938] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf9fae2f-8433-4c28-aa44-7ab3e22ef46c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.554543] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2058f16-6fc8-41a3-a149-f5c1f00f7ae2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.572045] env[61987]: DEBUG nova.compute.provider_tree [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1088.596662] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 8f9aab3e-96ee-44c6-8d25-706a0ff857c4] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1088.610432] env[61987]: DEBUG oslo_vmware.api [None req-b9c53053-20de-4495-8a51-051f14859ec1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062340, 'name': ReconfigVM_Task, 'duration_secs': 0.277396} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.613934] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9c53053-20de-4495-8a51-051f14859ec1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Reconfigured VM instance instance-0000006d to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1088.619317] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0059dab-bfb6-49c1-ab57-e0f4363ba8e5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.637848] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062341, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.639461] env[61987]: DEBUG oslo_vmware.api [None req-b9c53053-20de-4495-8a51-051f14859ec1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 1088.639461] env[61987]: value = "task-1062342" [ 1088.639461] env[61987]: _type = "Task" [ 1088.639461] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.656249] env[61987]: DEBUG oslo_vmware.api [None req-b9c53053-20de-4495-8a51-051f14859ec1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062342, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.727057] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c09f1308-0f02-432e-bad8-e993354e132a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.754423] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1088.754836] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f6f80961-031d-489d-ac84-e23a224e839e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.763436] env[61987]: DEBUG oslo_vmware.api [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1088.763436] env[61987]: value = "task-1062343" [ 1088.763436] env[61987]: _type = "Task" [ 1088.763436] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.772875] env[61987]: DEBUG oslo_vmware.api [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062343, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.075224] env[61987]: DEBUG nova.scheduler.client.report [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1089.104494] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 1961c2c7-4400-4559-9f7c-bec8e8e263b0] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1089.120082] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062341, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.751663} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.120325] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 5912b687-2150-4133-a3a1-44ce38b3c3cb/5912b687-2150-4133-a3a1-44ce38b3c3cb.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1089.120551] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1089.120810] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-28eac04d-9ef1-49a9-b0e6-3d5785bb77ae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.127284] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1089.127284] env[61987]: value = "task-1062344" [ 1089.127284] env[61987]: _type = "Task" [ 1089.127284] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.135575] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062344, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.148783] env[61987]: DEBUG oslo_vmware.api [None req-b9c53053-20de-4495-8a51-051f14859ec1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062342, 'name': ReconfigVM_Task, 'duration_secs': 0.161706} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.148882] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9c53053-20de-4495-8a51-051f14859ec1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234435', 'volume_id': '7c9a3b53-0803-45d3-9ebd-1ba92fed6944', 'name': 'volume-7c9a3b53-0803-45d3-9ebd-1ba92fed6944', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'bf1123bf-80e6-4be3-804f-e13b906ee44d', 'attached_at': '', 'detached_at': '', 'volume_id': '7c9a3b53-0803-45d3-9ebd-1ba92fed6944', 'serial': '7c9a3b53-0803-45d3-9ebd-1ba92fed6944'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1089.272985] env[61987]: DEBUG oslo_vmware.api [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062343, 'name': PowerOffVM_Task, 'duration_secs': 0.248148} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.273313] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1089.278727] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Reconfiguring VM instance instance-0000006e to detach disk 2002 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1089.278997] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-265d7368-dba4-4a7d-b048-e3d63fb531f7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.297035] env[61987]: DEBUG oslo_vmware.api [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1089.297035] env[61987]: value = "task-1062345" [ 1089.297035] env[61987]: _type = "Task" [ 1089.297035] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.304847] env[61987]: DEBUG oslo_vmware.api [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062345, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.580826] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.693s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.598905] env[61987]: INFO nova.scheduler.client.report [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Deleted allocations for instance 7be2179d-68f7-4103-a8c1-b6e5b88b8706 [ 1089.607044] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 6ed3f0e9-3c39-4406-9863-931a117961a8] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1089.638409] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062344, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.18893} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.638861] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1089.639595] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03bc226d-ae81-4b26-9f66-242bcd32afe5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.663972] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] 5912b687-2150-4133-a3a1-44ce38b3c3cb/5912b687-2150-4133-a3a1-44ce38b3c3cb.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1089.664494] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e23311d-2845-48db-9d58-9b8801a7f10f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.683556] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1089.683556] env[61987]: value = "task-1062346" [ 1089.683556] env[61987]: _type = "Task" [ 1089.683556] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.691451] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062346, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.692690] env[61987]: DEBUG nova.objects.instance [None req-b9c53053-20de-4495-8a51-051f14859ec1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lazy-loading 'flavor' on Instance uuid bf1123bf-80e6-4be3-804f-e13b906ee44d {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1089.807220] env[61987]: DEBUG oslo_vmware.api [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062345, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.106722] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a6da1be5-07b2-4182-a5e4-82c4cc844ee7 tempest-AttachInterfacesTestJSON-1224967408 tempest-AttachInterfacesTestJSON-1224967408-project-member] Lock "7be2179d-68f7-4103-a8c1-b6e5b88b8706" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.587s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.109701] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 98d895ee-1d1f-49cd-a055-fe6c8462a692] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1090.193490] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062346, 'name': ReconfigVM_Task, 'duration_secs': 0.250862} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.193958] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Reconfigured VM instance instance-00000072 to attach disk [datastore2] 5912b687-2150-4133-a3a1-44ce38b3c3cb/5912b687-2150-4133-a3a1-44ce38b3c3cb.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1090.196013] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-401c8a86-9962-40a9-9d08-dcdfda9f9d6e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.204098] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1090.204098] env[61987]: value = "task-1062347" [ 1090.204098] env[61987]: _type = "Task" [ 1090.204098] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.212550] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062347, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.307291] env[61987]: DEBUG oslo_vmware.api [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062345, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.612632] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 93706a54-1fc4-4aed-8807-ec2c3d3694b4] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1090.700029] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b9c53053-20de-4495-8a51-051f14859ec1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "bf1123bf-80e6-4be3-804f-e13b906ee44d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.241s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.714404] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062347, 'name': Rename_Task, 'duration_secs': 0.169582} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.714719] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1090.715015] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-617dbc1e-8cdd-4433-b2c2-3f58df3044a4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.721485] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1090.721485] env[61987]: value = "task-1062348" [ 1090.721485] env[61987]: _type = "Task" [ 1090.721485] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.728756] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062348, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.808235] env[61987]: DEBUG oslo_vmware.api [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062345, 'name': ReconfigVM_Task, 'duration_secs': 1.163379} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.808577] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Reconfigured VM instance instance-0000006e to detach disk 2002 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1090.808860] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1090.809135] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-247f9537-df82-43c5-8d28-a8558d83787e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.815103] env[61987]: DEBUG oslo_vmware.api [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1090.815103] env[61987]: value = "task-1062349" [ 1090.815103] env[61987]: _type = "Task" [ 1090.815103] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.823048] env[61987]: DEBUG oslo_vmware.api [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062349, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.116632] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: fe22fc1b-61fb-4088-b5fe-b84eb4ec4099] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1091.238726] env[61987]: DEBUG oslo_vmware.api [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062348, 'name': PowerOnVM_Task, 'duration_secs': 0.446231} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.239573] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1091.239805] env[61987]: INFO nova.compute.manager [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Took 7.54 seconds to spawn the instance on the hypervisor. [ 1091.240013] env[61987]: DEBUG nova.compute.manager [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1091.241233] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524eb885-2fa8-4260-bfa4-50a5488d1e2a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.327131] env[61987]: DEBUG oslo_vmware.api [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062349, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.623807] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 59d34772-26a5-4af2-82dc-da932e996745] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1091.727029] env[61987]: DEBUG oslo_concurrency.lockutils [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "bf1123bf-80e6-4be3-804f-e13b906ee44d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.727323] env[61987]: DEBUG oslo_concurrency.lockutils [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "bf1123bf-80e6-4be3-804f-e13b906ee44d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.727615] env[61987]: DEBUG oslo_concurrency.lockutils [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "bf1123bf-80e6-4be3-804f-e13b906ee44d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.727831] env[61987]: DEBUG oslo_concurrency.lockutils [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "bf1123bf-80e6-4be3-804f-e13b906ee44d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.728038] env[61987]: DEBUG oslo_concurrency.lockutils [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "bf1123bf-80e6-4be3-804f-e13b906ee44d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.730252] env[61987]: INFO nova.compute.manager [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Terminating instance [ 1091.759977] env[61987]: INFO nova.compute.manager [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Took 12.34 seconds to build instance. [ 1091.826258] env[61987]: DEBUG oslo_vmware.api [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062349, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.127646] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 1f29d5c1-bd01-47dc-9dcf-4e89791120d0] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1092.233914] env[61987]: DEBUG nova.compute.manager [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1092.234200] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1092.235111] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8098f51-33bc-46e0-af27-42656c68c3ca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.247273] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1092.247273] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-de098339-f6f9-41e0-9d2c-76ff4bc071e3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.254228] env[61987]: DEBUG oslo_vmware.api [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 1092.254228] env[61987]: value = "task-1062350" [ 1092.254228] env[61987]: _type = "Task" [ 1092.254228] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.262064] env[61987]: DEBUG oslo_concurrency.lockutils [None req-f5c7cad9-dd2a-45aa-aef4-45a913df1900 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "5912b687-2150-4133-a3a1-44ce38b3c3cb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.849s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.262317] env[61987]: DEBUG oslo_vmware.api [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062350, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.326620] env[61987]: DEBUG oslo_vmware.api [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062349, 'name': PowerOnVM_Task, 'duration_secs': 1.040455} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.326903] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1092.327160] env[61987]: DEBUG nova.compute.manager [None req-a799c12e-583b-4e5d-a824-d4b8cba2fa9f tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1092.328105] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02d7c38f-5666-4812-ac6b-6981427d5be0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.630777] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 4fd8fc17-fb13-42e7-9879-9f120fdcb330] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1092.765732] env[61987]: DEBUG oslo_vmware.api [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062350, 'name': PowerOffVM_Task, 'duration_secs': 0.178572} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.765732] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1092.765895] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1092.766532] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-044e2335-7c0e-446f-89ed-e1b3cbe01f81 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.839174] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1092.839174] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1092.839174] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Deleting the datastore file [datastore1] bf1123bf-80e6-4be3-804f-e13b906ee44d {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1092.840234] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3663a471-35c9-4916-82b0-48a8c8f985f4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.848716] env[61987]: DEBUG oslo_vmware.api [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for the task: (returnval){ [ 1092.848716] env[61987]: value = "task-1062352" [ 1092.848716] env[61987]: _type = "Task" [ 1092.848716] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.857217] env[61987]: DEBUG oslo_vmware.api [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062352, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.135026] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 84a4a69b-04db-4546-ba89-9039e382a0c4] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1093.360238] env[61987]: DEBUG oslo_vmware.api [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Task: {'id': task-1062352, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134021} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.360588] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1093.360863] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1093.361115] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1093.361367] env[61987]: INFO nova.compute.manager [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1093.361697] env[61987]: DEBUG oslo.service.loopingcall [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1093.361958] env[61987]: DEBUG nova.compute.manager [-] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1093.362126] env[61987]: DEBUG nova.network.neutron [-] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1093.637991] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 10354ba0-5a64-4174-9c61-152d54a5dfcc] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1093.689300] env[61987]: DEBUG nova.compute.manager [req-077342af-f2a5-46de-9d29-67bac82b2a00 req-1d7a9b22-ac89-4501-a96a-4959655b4b81 service nova] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Received event network-changed-a8217ebd-b52f-4664-a191-dc7f3513a366 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1093.690156] env[61987]: DEBUG nova.compute.manager [req-077342af-f2a5-46de-9d29-67bac82b2a00 req-1d7a9b22-ac89-4501-a96a-4959655b4b81 service nova] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Refreshing instance network info cache due to event network-changed-a8217ebd-b52f-4664-a191-dc7f3513a366. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1093.690156] env[61987]: DEBUG oslo_concurrency.lockutils [req-077342af-f2a5-46de-9d29-67bac82b2a00 req-1d7a9b22-ac89-4501-a96a-4959655b4b81 service nova] Acquiring lock "refresh_cache-5912b687-2150-4133-a3a1-44ce38b3c3cb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.690156] env[61987]: DEBUG oslo_concurrency.lockutils [req-077342af-f2a5-46de-9d29-67bac82b2a00 req-1d7a9b22-ac89-4501-a96a-4959655b4b81 service nova] Acquired lock "refresh_cache-5912b687-2150-4133-a3a1-44ce38b3c3cb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.690478] env[61987]: DEBUG nova.network.neutron [req-077342af-f2a5-46de-9d29-67bac82b2a00 req-1d7a9b22-ac89-4501-a96a-4959655b4b81 service nova] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Refreshing network info cache for port a8217ebd-b52f-4664-a191-dc7f3513a366 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1094.095859] env[61987]: DEBUG nova.compute.manager [req-38081583-b1ea-4eeb-b015-ba8118840c1b req-d8f7105e-c679-43b6-a6c0-c2764ffd6e7f service nova] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Received event network-vif-deleted-a46b1587-a8f9-4552-8d05-57882397c3f5 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1094.096098] env[61987]: INFO nova.compute.manager [req-38081583-b1ea-4eeb-b015-ba8118840c1b req-d8f7105e-c679-43b6-a6c0-c2764ffd6e7f service nova] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Neutron deleted interface a46b1587-a8f9-4552-8d05-57882397c3f5; detaching it from the instance and deleting it from the info cache [ 1094.096305] env[61987]: DEBUG nova.network.neutron [req-38081583-b1ea-4eeb-b015-ba8118840c1b req-d8f7105e-c679-43b6-a6c0-c2764ffd6e7f service nova] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.141444] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 30c4fff0-c77e-4c11-aced-a040003a6b10] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1094.491331] env[61987]: DEBUG nova.network.neutron [req-077342af-f2a5-46de-9d29-67bac82b2a00 req-1d7a9b22-ac89-4501-a96a-4959655b4b81 service nova] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Updated VIF entry in instance network info cache for port a8217ebd-b52f-4664-a191-dc7f3513a366. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1094.491720] env[61987]: DEBUG nova.network.neutron [req-077342af-f2a5-46de-9d29-67bac82b2a00 req-1d7a9b22-ac89-4501-a96a-4959655b4b81 service nova] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Updating instance_info_cache with network_info: [{"id": "a8217ebd-b52f-4664-a191-dc7f3513a366", "address": "fa:16:3e:49:1b:98", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8217ebd-b5", "ovs_interfaceid": "a8217ebd-b52f-4664-a191-dc7f3513a366", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.563223] env[61987]: DEBUG nova.network.neutron [-] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.604983] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2dcbcc56-da15-4579-b80e-d03d7c7bb129 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.615127] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47fdd263-9ea4-450f-9bc3-015e8d5e1a24 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.648958] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 93a27dea-f28a-4cbf-ac37-ff41763ae18d] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1094.651487] env[61987]: DEBUG nova.compute.manager [req-38081583-b1ea-4eeb-b015-ba8118840c1b req-d8f7105e-c679-43b6-a6c0-c2764ffd6e7f service nova] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Detach interface failed, port_id=a46b1587-a8f9-4552-8d05-57882397c3f5, reason: Instance bf1123bf-80e6-4be3-804f-e13b906ee44d could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 1094.994378] env[61987]: DEBUG oslo_concurrency.lockutils [req-077342af-f2a5-46de-9d29-67bac82b2a00 req-1d7a9b22-ac89-4501-a96a-4959655b4b81 service nova] Releasing lock "refresh_cache-5912b687-2150-4133-a3a1-44ce38b3c3cb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.066324] env[61987]: INFO nova.compute.manager [-] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Took 1.70 seconds to deallocate network for instance. [ 1095.152206] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 1c63bf62-3f70-42b9-b6d7-41336d20f6c8] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1095.573677] env[61987]: DEBUG oslo_concurrency.lockutils [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.573991] env[61987]: DEBUG oslo_concurrency.lockutils [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.574459] env[61987]: DEBUG nova.objects.instance [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lazy-loading 'resources' on Instance uuid bf1123bf-80e6-4be3-804f-e13b906ee44d {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1095.655328] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 8e68c5a2-6587-45b6-9a76-96b129a0a665] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1096.124373] env[61987]: DEBUG nova.compute.manager [req-6ff09225-0e5b-42c7-9352-20527a77adb1 req-9d550d7b-02b6-4956-bef3-cb800b98cdfd service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Received event network-changed-d96dede6-c439-4a14-a4d1-62c44e03ea11 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1096.124613] env[61987]: DEBUG nova.compute.manager [req-6ff09225-0e5b-42c7-9352-20527a77adb1 req-9d550d7b-02b6-4956-bef3-cb800b98cdfd service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Refreshing instance network info cache due to event network-changed-d96dede6-c439-4a14-a4d1-62c44e03ea11. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1096.125112] env[61987]: DEBUG oslo_concurrency.lockutils [req-6ff09225-0e5b-42c7-9352-20527a77adb1 req-9d550d7b-02b6-4956-bef3-cb800b98cdfd service nova] Acquiring lock "refresh_cache-7002962c-76ed-4a8d-be2a-b2b118847fea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.125112] env[61987]: DEBUG oslo_concurrency.lockutils [req-6ff09225-0e5b-42c7-9352-20527a77adb1 req-9d550d7b-02b6-4956-bef3-cb800b98cdfd service nova] Acquired lock "refresh_cache-7002962c-76ed-4a8d-be2a-b2b118847fea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.125334] env[61987]: DEBUG nova.network.neutron [req-6ff09225-0e5b-42c7-9352-20527a77adb1 req-9d550d7b-02b6-4956-bef3-cb800b98cdfd service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Refreshing network info cache for port d96dede6-c439-4a14-a4d1-62c44e03ea11 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1096.158691] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: d566ea53-958b-4a35-989c-771180d95bb4] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1096.188715] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b9a9914-07a4-4d88-8091-06ee7ff875c6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.196772] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c954b579-ad62-4cc6-91a8-4e3d0655904e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.226641] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e455b0b-309a-45cd-9d65-4582e80ccb93 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.234033] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7934261a-0bcb-4203-a996-c48e3cd347e2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.248672] env[61987]: DEBUG nova.compute.provider_tree [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1096.663872] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: f937a5ec-a1d7-41d8-b998-fa18f545b304] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1096.752978] env[61987]: DEBUG nova.scheduler.client.report [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1096.868454] env[61987]: DEBUG nova.network.neutron [req-6ff09225-0e5b-42c7-9352-20527a77adb1 req-9d550d7b-02b6-4956-bef3-cb800b98cdfd service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Updated VIF entry in instance network info cache for port d96dede6-c439-4a14-a4d1-62c44e03ea11. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1096.868454] env[61987]: DEBUG nova.network.neutron [req-6ff09225-0e5b-42c7-9352-20527a77adb1 req-9d550d7b-02b6-4956-bef3-cb800b98cdfd service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Updating instance_info_cache with network_info: [{"id": "d96dede6-c439-4a14-a4d1-62c44e03ea11", "address": "fa:16:3e:d2:e0:b1", "network": {"id": "4ed4fbdc-05b0-44f8-a3d5-b7b288abd5e0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2067313775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a234a62da0e498fbe613fbcaaec3201", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd96dede6-c4", "ovs_interfaceid": "d96dede6-c439-4a14-a4d1-62c44e03ea11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.167889] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 8cb6f7b4-611d-4f10-8ee0-b2c044aec5c2] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1097.260580] env[61987]: DEBUG oslo_concurrency.lockutils [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.686s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.283041] env[61987]: INFO nova.scheduler.client.report [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Deleted allocations for instance bf1123bf-80e6-4be3-804f-e13b906ee44d [ 1097.371934] env[61987]: DEBUG oslo_concurrency.lockutils [req-6ff09225-0e5b-42c7-9352-20527a77adb1 req-9d550d7b-02b6-4956-bef3-cb800b98cdfd service nova] Releasing lock "refresh_cache-7002962c-76ed-4a8d-be2a-b2b118847fea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.372207] env[61987]: DEBUG nova.compute.manager [req-6ff09225-0e5b-42c7-9352-20527a77adb1 req-9d550d7b-02b6-4956-bef3-cb800b98cdfd service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Received event network-changed-d96dede6-c439-4a14-a4d1-62c44e03ea11 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1097.372393] env[61987]: DEBUG nova.compute.manager [req-6ff09225-0e5b-42c7-9352-20527a77adb1 req-9d550d7b-02b6-4956-bef3-cb800b98cdfd service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Refreshing instance network info cache due to event network-changed-d96dede6-c439-4a14-a4d1-62c44e03ea11. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1097.372617] env[61987]: DEBUG oslo_concurrency.lockutils [req-6ff09225-0e5b-42c7-9352-20527a77adb1 req-9d550d7b-02b6-4956-bef3-cb800b98cdfd service nova] Acquiring lock "refresh_cache-7002962c-76ed-4a8d-be2a-b2b118847fea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1097.372769] env[61987]: DEBUG oslo_concurrency.lockutils [req-6ff09225-0e5b-42c7-9352-20527a77adb1 req-9d550d7b-02b6-4956-bef3-cb800b98cdfd service nova] Acquired lock "refresh_cache-7002962c-76ed-4a8d-be2a-b2b118847fea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.372939] env[61987]: DEBUG nova.network.neutron [req-6ff09225-0e5b-42c7-9352-20527a77adb1 req-9d550d7b-02b6-4956-bef3-cb800b98cdfd service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Refreshing network info cache for port d96dede6-c439-4a14-a4d1-62c44e03ea11 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1097.673019] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 95641d0b-970c-4b94-8568-5c46d0808345] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1097.790752] env[61987]: DEBUG oslo_concurrency.lockutils [None req-502b90d9-29dc-43b0-8867-6624ad6ee5c1 tempest-AttachVolumeNegativeTest-583696460 tempest-AttachVolumeNegativeTest-583696460-project-member] Lock "bf1123bf-80e6-4be3-804f-e13b906ee44d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.063s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.101836] env[61987]: DEBUG nova.network.neutron [req-6ff09225-0e5b-42c7-9352-20527a77adb1 req-9d550d7b-02b6-4956-bef3-cb800b98cdfd service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Updated VIF entry in instance network info cache for port d96dede6-c439-4a14-a4d1-62c44e03ea11. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1098.102924] env[61987]: DEBUG nova.network.neutron [req-6ff09225-0e5b-42c7-9352-20527a77adb1 req-9d550d7b-02b6-4956-bef3-cb800b98cdfd service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Updating instance_info_cache with network_info: [{"id": "d96dede6-c439-4a14-a4d1-62c44e03ea11", "address": "fa:16:3e:d2:e0:b1", "network": {"id": "4ed4fbdc-05b0-44f8-a3d5-b7b288abd5e0", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-2067313775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.238", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a234a62da0e498fbe613fbcaaec3201", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ddfb706a-add1-4e16-9ac4-d20b16a1df6d", "external-id": "nsx-vlan-transportzone-820", "segmentation_id": 820, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd96dede6-c4", "ovs_interfaceid": "d96dede6-c439-4a14-a4d1-62c44e03ea11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1098.176462] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: ad40937f-d5a9-4708-8b81-06087d38d765] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1098.604786] env[61987]: DEBUG oslo_concurrency.lockutils [req-6ff09225-0e5b-42c7-9352-20527a77adb1 req-9d550d7b-02b6-4956-bef3-cb800b98cdfd service nova] Releasing lock "refresh_cache-7002962c-76ed-4a8d-be2a-b2b118847fea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.683744] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: d038391c-2b32-455c-85d5-68221e7bd065] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1099.187793] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 4edd0159-6db1-41bd-a616-31a34e984059] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1099.690827] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 80f13fca-2df1-4aa0-96f1-b61a60c4fdc6] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1100.194887] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 51a36102-795e-47b7-b96a-857e54dc703e] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1100.697952] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1100.698140] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Cleaning up deleted instances with incomplete migration {{(pid=61987) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11739}} [ 1101.201159] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1103.698919] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1103.699358] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1104.206092] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1104.206092] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Starting heal instance info cache {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10401}} [ 1106.265540] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a8d9a3f-2968-4515-b2d4-d27706e32998 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "76707afe-29ae-4257-a237-ba5a47b2511b" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.265851] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a8d9a3f-2968-4515-b2d4-d27706e32998 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "76707afe-29ae-4257-a237-ba5a47b2511b" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.751679] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "refresh_cache-4d15848d-5265-4897-bfa7-f3ad4c02ed83" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1106.751842] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquired lock "refresh_cache-4d15848d-5265-4897-bfa7-f3ad4c02ed83" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.752024] env[61987]: DEBUG nova.network.neutron [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Forcefully refreshing network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1106.769239] env[61987]: DEBUG nova.compute.utils [None req-0a8d9a3f-2968-4515-b2d4-d27706e32998 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1107.273397] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a8d9a3f-2968-4515-b2d4-d27706e32998 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "76707afe-29ae-4257-a237-ba5a47b2511b" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.971893] env[61987]: DEBUG nova.network.neutron [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Updating instance_info_cache with network_info: [{"id": "203d81e4-dbb8-4685-a463-9de48b835ff7", "address": "fa:16:3e:5c:c5:2c", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap203d81e4-db", "ovs_interfaceid": "203d81e4-dbb8-4685-a463-9de48b835ff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.336855] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a8d9a3f-2968-4515-b2d4-d27706e32998 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "76707afe-29ae-4257-a237-ba5a47b2511b" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.337231] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a8d9a3f-2968-4515-b2d4-d27706e32998 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "76707afe-29ae-4257-a237-ba5a47b2511b" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.337429] env[61987]: INFO nova.compute.manager [None req-0a8d9a3f-2968-4515-b2d4-d27706e32998 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Attaching volume 0f768543-5a0b-4e1d-887c-6cb5f60afe63 to /dev/sdb [ 1108.368981] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d282f85-cd39-46aa-afba-e4c954b610e2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.377039] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c3bda46-5faf-4397-b5b8-f7add256ac3a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.391150] env[61987]: DEBUG nova.virt.block_device [None req-0a8d9a3f-2968-4515-b2d4-d27706e32998 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Updating existing volume attachment record: 8c67082f-98c6-49f4-80a0-c6349f74af0e {{(pid=61987) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1108.474862] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Releasing lock "refresh_cache-4d15848d-5265-4897-bfa7-f3ad4c02ed83" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.475206] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Updated the network info_cache for instance {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10472}} [ 1108.475454] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1108.475654] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1108.475828] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1108.475986] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1108.476149] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1108.476312] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1108.476446] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61987) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11020}} [ 1108.476588] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1108.980801] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.981052] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.981233] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.981395] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61987) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1108.982328] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faace97b-99c1-4619-b1cb-ce454928f118 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.992496] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81183429-4a89-4edd-8c28-6554246de9f1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.007224] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b9100c7-4d63-4b38-8d82-b9acecd5020e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.013758] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4350993-9861-418b-8a4a-54b55cc1ccf5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.042330] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180098MB free_disk=179GB free_vcpus=48 pci_devices=None {{(pid=61987) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1109.042490] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.042682] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.072242] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 4d15848d-5265-4897-bfa7-f3ad4c02ed83 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1110.072513] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1110.072566] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 473e0b73-6fcb-44c1-abf7-47d30c58af1b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1110.072649] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 7002962c-76ed-4a8d-be2a-b2b118847fea actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1110.072768] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 76707afe-29ae-4257-a237-ba5a47b2511b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1110.072886] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance f030044b-bc6b-478c-8f63-d519aa76203c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1110.073015] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 5912b687-2150-4133-a3a1-44ce38b3c3cb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1110.073220] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1110.073360] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1856MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1110.149813] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a82e26-1cc3-4db8-8b40-968bcbbc4a83 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.157108] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5110c675-a9f0-4acb-81bd-facd95593d4d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.186052] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a41cd9-dc44-423d-a533-8dd56bbc80c7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.193270] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7b0b40f-d38f-4322-b815-ee00f833bf8d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.206469] env[61987]: DEBUG nova.compute.provider_tree [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1110.710510] env[61987]: DEBUG nova.scheduler.client.report [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1111.217935] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61987) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1111.218237] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.175s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.936614] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a8d9a3f-2968-4515-b2d4-d27706e32998 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Volume attach. Driver type: vmdk {{(pid=61987) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1112.936964] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a8d9a3f-2968-4515-b2d4-d27706e32998 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234443', 'volume_id': '0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'name': 'volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '76707afe-29ae-4257-a237-ba5a47b2511b', 'attached_at': '', 'detached_at': '', 'volume_id': '0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'serial': '0f768543-5a0b-4e1d-887c-6cb5f60afe63'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1112.937794] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1de2788e-1752-4c89-903a-81c4641c1078 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.955439] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e983d194-3e21-4b0d-a45f-98d004fb598d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.979205] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a8d9a3f-2968-4515-b2d4-d27706e32998 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63/volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1112.979436] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34e6af09-b6f4-4e26-91a0-a49bf2d09423 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.997337] env[61987]: DEBUG oslo_vmware.api [None req-0a8d9a3f-2968-4515-b2d4-d27706e32998 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1112.997337] env[61987]: value = "task-1062370" [ 1112.997337] env[61987]: _type = "Task" [ 1112.997337] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.005572] env[61987]: DEBUG oslo_vmware.api [None req-0a8d9a3f-2968-4515-b2d4-d27706e32998 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062370, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.243842] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._sync_power_states {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1113.507639] env[61987]: DEBUG oslo_vmware.api [None req-0a8d9a3f-2968-4515-b2d4-d27706e32998 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062370, 'name': ReconfigVM_Task, 'duration_secs': 0.323881} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.507941] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a8d9a3f-2968-4515-b2d4-d27706e32998 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Reconfigured VM instance instance-00000070 to attach disk [datastore2] volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63/volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1113.512690] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1345947e-9240-4780-82e6-eeb5d37284aa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.528186] env[61987]: DEBUG oslo_vmware.api [None req-0a8d9a3f-2968-4515-b2d4-d27706e32998 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1113.528186] env[61987]: value = "task-1062371" [ 1113.528186] env[61987]: _type = "Task" [ 1113.528186] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.535951] env[61987]: DEBUG oslo_vmware.api [None req-0a8d9a3f-2968-4515-b2d4-d27706e32998 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062371, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.748466] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Getting list of instances from cluster (obj){ [ 1113.748466] env[61987]: value = "domain-c8" [ 1113.748466] env[61987]: _type = "ClusterComputeResource" [ 1113.748466] env[61987]: } {{(pid=61987) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 1113.749848] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4220a985-051b-4798-9eac-d0d6d72b0c6f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.765795] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Got total of 7 instances {{(pid=61987) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 1113.765974] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Triggering sync for uuid 4d15848d-5265-4897-bfa7-f3ad4c02ed83 {{(pid=61987) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10811}} [ 1113.766241] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Triggering sync for uuid aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8 {{(pid=61987) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10811}} [ 1113.766419] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Triggering sync for uuid 473e0b73-6fcb-44c1-abf7-47d30c58af1b {{(pid=61987) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10811}} [ 1113.766580] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Triggering sync for uuid 7002962c-76ed-4a8d-be2a-b2b118847fea {{(pid=61987) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10811}} [ 1113.766732] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Triggering sync for uuid 76707afe-29ae-4257-a237-ba5a47b2511b {{(pid=61987) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10811}} [ 1113.766917] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Triggering sync for uuid f030044b-bc6b-478c-8f63-d519aa76203c {{(pid=61987) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10811}} [ 1113.767097] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Triggering sync for uuid 5912b687-2150-4133-a3a1-44ce38b3c3cb {{(pid=61987) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10811}} [ 1113.767454] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "4d15848d-5265-4897-bfa7-f3ad4c02ed83" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.767681] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "4d15848d-5265-4897-bfa7-f3ad4c02ed83" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.767988] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.768167] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.768409] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "473e0b73-6fcb-44c1-abf7-47d30c58af1b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.768599] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "473e0b73-6fcb-44c1-abf7-47d30c58af1b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.768835] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "7002962c-76ed-4a8d-be2a-b2b118847fea" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.769034] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "7002962c-76ed-4a8d-be2a-b2b118847fea" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.769338] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "76707afe-29ae-4257-a237-ba5a47b2511b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.769556] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "f030044b-bc6b-478c-8f63-d519aa76203c" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.769744] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "f030044b-bc6b-478c-8f63-d519aa76203c" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.769985] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "5912b687-2150-4133-a3a1-44ce38b3c3cb" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.770195] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "5912b687-2150-4133-a3a1-44ce38b3c3cb" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1113.771014] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b1b342c-9cca-4b81-bd27-71f99bd7c33f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.774152] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f7abe22-5676-4c04-83fd-e91187cd2fb5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.776908] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b9ea8e6-c4be-4f55-af7c-1ee877a6925c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.779687] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59406b39-2cf0-4e74-ad26-786f54f3a2f4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.782245] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1e3f1bf-7812-4541-aeee-11e836414ddb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.784917] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9184b827-f466-4b4f-bf1f-913915aa45f8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.038318] env[61987]: DEBUG oslo_vmware.api [None req-0a8d9a3f-2968-4515-b2d4-d27706e32998 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062371, 'name': ReconfigVM_Task, 'duration_secs': 0.137853} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.038837] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a8d9a3f-2968-4515-b2d4-d27706e32998 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234443', 'volume_id': '0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'name': 'volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '76707afe-29ae-4257-a237-ba5a47b2511b', 'attached_at': '', 'detached_at': '', 'volume_id': '0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'serial': '0f768543-5a0b-4e1d-887c-6cb5f60afe63'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1114.305193] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.537s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.305677] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "473e0b73-6fcb-44c1-abf7-47d30c58af1b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.537s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.306064] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "f030044b-bc6b-478c-8f63-d519aa76203c" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.536s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.313960] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "5912b687-2150-4133-a3a1-44ce38b3c3cb" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.544s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.314303] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "7002962c-76ed-4a8d-be2a-b2b118847fea" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.545s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.314632] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "4d15848d-5265-4897-bfa7-f3ad4c02ed83" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.547s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.075107] env[61987]: DEBUG nova.objects.instance [None req-0a8d9a3f-2968-4515-b2d4-d27706e32998 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lazy-loading 'flavor' on Instance uuid 76707afe-29ae-4257-a237-ba5a47b2511b {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1115.582509] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0a8d9a3f-2968-4515-b2d4-d27706e32998 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "76707afe-29ae-4257-a237-ba5a47b2511b" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.245s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.583521] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "76707afe-29ae-4257-a237-ba5a47b2511b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 1.814s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.584489] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2348ea3b-065f-4ee5-b249-0045866134a8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.094574] env[61987]: INFO nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] During sync_power_state the instance has a pending task (rebuilding). Skip. [ 1116.094864] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "76707afe-29ae-4257-a237-ba5a47b2511b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.511s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.115546] env[61987]: INFO nova.compute.manager [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Rebuilding instance [ 1116.154433] env[61987]: DEBUG nova.compute.manager [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1116.155285] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ae86040-9f65-49f5-b927-1a52bfcf8b49 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.610611] env[61987]: DEBUG oslo_concurrency.lockutils [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Acquiring lock "2ba77d32-d777-448e-8fa6-04920950f0f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.610880] env[61987]: DEBUG oslo_concurrency.lockutils [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Lock "2ba77d32-d777-448e-8fa6-04920950f0f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.113384] env[61987]: DEBUG nova.compute.manager [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1117.169163] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1117.169492] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-861d0c13-3230-4a13-a49b-70adc8ed039d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.177713] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1117.177713] env[61987]: value = "task-1062372" [ 1117.177713] env[61987]: _type = "Task" [ 1117.177713] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.187018] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062372, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.635994] env[61987]: DEBUG oslo_concurrency.lockutils [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.636306] env[61987]: DEBUG oslo_concurrency.lockutils [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.637940] env[61987]: INFO nova.compute.claims [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1117.687580] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062372, 'name': PowerOffVM_Task, 'duration_secs': 0.193551} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.687872] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1117.740674] env[61987]: INFO nova.compute.manager [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Detaching volume 0f768543-5a0b-4e1d-887c-6cb5f60afe63 [ 1117.772035] env[61987]: INFO nova.virt.block_device [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Attempting to driver detach volume 0f768543-5a0b-4e1d-887c-6cb5f60afe63 from mountpoint /dev/sdb [ 1117.772366] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Volume detach. Driver type: vmdk {{(pid=61987) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1117.772589] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234443', 'volume_id': '0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'name': 'volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '76707afe-29ae-4257-a237-ba5a47b2511b', 'attached_at': '', 'detached_at': '', 'volume_id': '0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'serial': '0f768543-5a0b-4e1d-887c-6cb5f60afe63'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1117.773530] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eba5ab7-fe0f-4dd1-9f66-144a838c36e3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.794880] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d495a0fb-166b-45f3-a57d-0b34d36ff4b5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.801993] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32fd3d55-629a-4ee7-b472-3d5db60e3bf2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.822140] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae070ef-a40c-47df-9bf1-09196c70a31f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.836399] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] The volume has not been displaced from its original location: [datastore2] volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63/volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63.vmdk. No consolidation needed. {{(pid=61987) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1117.841856] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Reconfiguring VM instance instance-00000070 to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1117.842016] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90087481-907a-4df3-b447-57e298f45782 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.860248] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1117.860248] env[61987]: value = "task-1062373" [ 1117.860248] env[61987]: _type = "Task" [ 1117.860248] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.868035] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062373, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.371928] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062373, 'name': ReconfigVM_Task, 'duration_secs': 0.218891} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.372347] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Reconfigured VM instance instance-00000070 to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1118.376855] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-000bae5c-0515-404d-9679-7fd9f8de582c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.392595] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1118.392595] env[61987]: value = "task-1062374" [ 1118.392595] env[61987]: _type = "Task" [ 1118.392595] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.402777] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062374, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.752912] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e462304a-ce71-444c-9f1f-846cf64b18da {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.760868] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff898253-163c-42c5-a277-6bf8f212bf26 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.790813] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e4cfea0-df68-438c-a4e6-f3d9f67490ef {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.798838] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e97c76d8-0826-4e7e-be6f-67417cbb6e0d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.812405] env[61987]: DEBUG nova.compute.provider_tree [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1118.902654] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062374, 'name': ReconfigVM_Task, 'duration_secs': 0.198469} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.902894] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234443', 'volume_id': '0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'name': 'volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '76707afe-29ae-4257-a237-ba5a47b2511b', 'attached_at': '', 'detached_at': '', 'volume_id': '0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'serial': '0f768543-5a0b-4e1d-887c-6cb5f60afe63'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1119.315778] env[61987]: DEBUG nova.scheduler.client.report [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1119.821977] env[61987]: DEBUG oslo_concurrency.lockutils [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.185s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.821977] env[61987]: DEBUG nova.compute.manager [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1119.950174] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1119.950499] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3b61945d-e9a9-4ae9-8471-5c0a414097c8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.957668] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1119.957668] env[61987]: value = "task-1062375" [ 1119.957668] env[61987]: _type = "Task" [ 1119.957668] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.965484] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062375, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.328613] env[61987]: DEBUG nova.compute.utils [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1120.330711] env[61987]: DEBUG nova.compute.manager [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1120.330711] env[61987]: DEBUG nova.network.neutron [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1120.369398] env[61987]: DEBUG nova.policy [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '01240167b477416ba0787a7ca888cd70', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '931187d9051349cfaa9cfeeb2ddf032c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 1120.467639] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] VM already powered off {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1120.467888] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Volume detach. Driver type: vmdk {{(pid=61987) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1120.468111] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234443', 'volume_id': '0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'name': 'volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '76707afe-29ae-4257-a237-ba5a47b2511b', 'attached_at': '', 'detached_at': '', 'volume_id': '0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'serial': '0f768543-5a0b-4e1d-887c-6cb5f60afe63'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1120.468884] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-503cb22c-3d4b-4a5d-8067-e8d40261a19f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.486693] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74aab7d4-ae01-447d-b235-f08706a1e139 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.492579] env[61987]: WARNING nova.virt.vmwareapi.driver [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1120.492786] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1120.493523] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92fad6bb-1de3-4707-8b7b-8481e1c4acd1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.499709] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1120.499925] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6cbe0506-a452-4c82-8d93-39ee2dc204fb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.651418] env[61987]: DEBUG nova.network.neutron [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Successfully created port: 2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1120.833739] env[61987]: DEBUG nova.compute.manager [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1120.953772] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1120.954017] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1120.954223] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Deleting the datastore file [datastore1] 76707afe-29ae-4257-a237-ba5a47b2511b {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1120.954492] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d3c984ef-2d93-42c3-ba01-8845d739649a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.960830] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1120.960830] env[61987]: value = "task-1062377" [ 1120.960830] env[61987]: _type = "Task" [ 1120.960830] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.968747] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062377, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.339078] env[61987]: INFO nova.virt.block_device [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Booting with volume 4ab1a236-de8d-4578-8274-89bc106faa2e at /dev/sda [ 1121.373096] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a61b0832-6b82-4a1f-9937-b4a590b05ffb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.381712] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c89f4b25-789a-44f5-9374-295aa8dfaed5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.408392] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f541caa-ec02-4046-9db7-0ab22e65c3a8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.415764] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-045d8dfc-6c7f-4117-b98f-5a772c78210f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.442877] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7734b999-ca2f-4643-bfec-8d12042c8d43 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.448470] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f7c1d9-03c5-46a4-8f6a-db11e52b2c25 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.460600] env[61987]: DEBUG nova.virt.block_device [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Updating existing volume attachment record: 79215acc-dd55-4305-8076-ecbf5a58971a {{(pid=61987) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1121.469851] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062377, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127669} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.470120] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1121.470409] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1121.470595] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1121.975840] env[61987]: INFO nova.virt.block_device [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Booting with volume 0f768543-5a0b-4e1d-887c-6cb5f60afe63 at /dev/sdb [ 1122.038182] env[61987]: DEBUG nova.compute.manager [req-b0b1b421-bd1f-49b7-bc7e-b8a611b22e97 req-cdefc9ed-fdfe-4802-8679-ee43c706c3c1 service nova] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Received event network-vif-plugged-2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1122.038298] env[61987]: DEBUG oslo_concurrency.lockutils [req-b0b1b421-bd1f-49b7-bc7e-b8a611b22e97 req-cdefc9ed-fdfe-4802-8679-ee43c706c3c1 service nova] Acquiring lock "2ba77d32-d777-448e-8fa6-04920950f0f6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.038526] env[61987]: DEBUG oslo_concurrency.lockutils [req-b0b1b421-bd1f-49b7-bc7e-b8a611b22e97 req-cdefc9ed-fdfe-4802-8679-ee43c706c3c1 service nova] Lock "2ba77d32-d777-448e-8fa6-04920950f0f6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.038707] env[61987]: DEBUG oslo_concurrency.lockutils [req-b0b1b421-bd1f-49b7-bc7e-b8a611b22e97 req-cdefc9ed-fdfe-4802-8679-ee43c706c3c1 service nova] Lock "2ba77d32-d777-448e-8fa6-04920950f0f6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.038880] env[61987]: DEBUG nova.compute.manager [req-b0b1b421-bd1f-49b7-bc7e-b8a611b22e97 req-cdefc9ed-fdfe-4802-8679-ee43c706c3c1 service nova] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] No waiting events found dispatching network-vif-plugged-2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1122.039276] env[61987]: WARNING nova.compute.manager [req-b0b1b421-bd1f-49b7-bc7e-b8a611b22e97 req-cdefc9ed-fdfe-4802-8679-ee43c706c3c1 service nova] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Received unexpected event network-vif-plugged-2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a for instance with vm_state building and task_state block_device_mapping. [ 1122.069648] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fdf8731c-eee7-488c-b97f-c8a89b8c36cf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.079687] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-298eb606-b8f6-4fba-ae98-d18a8414256d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.105758] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-60006a90-28c8-45a1-a60b-df4cb9f62d47 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.113224] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc9d970-6344-4c15-a8f1-5ead4cb86e65 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.125062] env[61987]: DEBUG nova.network.neutron [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Successfully updated port: 2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1122.142085] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4742afd-11c4-45f2-b131-28088c25932e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.148677] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b52abbf-bd98-402c-87c9-7ffc14b95c7a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.161845] env[61987]: DEBUG nova.virt.block_device [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Updating existing volume attachment record: 7f5bb1a0-8feb-4d80-9162-5b52924d8439 {{(pid=61987) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1122.636917] env[61987]: DEBUG oslo_concurrency.lockutils [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Acquiring lock "refresh_cache-2ba77d32-d777-448e-8fa6-04920950f0f6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.637228] env[61987]: DEBUG oslo_concurrency.lockutils [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Acquired lock "refresh_cache-2ba77d32-d777-448e-8fa6-04920950f0f6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.637450] env[61987]: DEBUG nova.network.neutron [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1123.194080] env[61987]: DEBUG nova.network.neutron [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1123.319708] env[61987]: DEBUG nova.network.neutron [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Updating instance_info_cache with network_info: [{"id": "2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a", "address": "fa:16:3e:99:ca:50", "network": {"id": "b42b6e20-733c-4c80-83c3-06665b85ade7", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-232939311-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931187d9051349cfaa9cfeeb2ddf032c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cf009f7-22", "ovs_interfaceid": "2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.542559] env[61987]: DEBUG nova.compute.manager [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1123.543174] env[61987]: DEBUG nova.virt.hardware [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1123.543407] env[61987]: DEBUG nova.virt.hardware [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1123.543576] env[61987]: DEBUG nova.virt.hardware [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1123.543770] env[61987]: DEBUG nova.virt.hardware [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1123.543924] env[61987]: DEBUG nova.virt.hardware [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1123.544105] env[61987]: DEBUG nova.virt.hardware [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1123.544324] env[61987]: DEBUG nova.virt.hardware [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1123.544574] env[61987]: DEBUG nova.virt.hardware [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1123.544661] env[61987]: DEBUG nova.virt.hardware [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1123.544829] env[61987]: DEBUG nova.virt.hardware [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1123.545021] env[61987]: DEBUG nova.virt.hardware [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1123.545914] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2734aa3-ee87-4272-9bf1-a915aafb247e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.554164] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f6168dd-db03-4c13-a0ee-d94b30115ee0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.822669] env[61987]: DEBUG oslo_concurrency.lockutils [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Releasing lock "refresh_cache-2ba77d32-d777-448e-8fa6-04920950f0f6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1123.822990] env[61987]: DEBUG nova.compute.manager [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Instance network_info: |[{"id": "2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a", "address": "fa:16:3e:99:ca:50", "network": {"id": "b42b6e20-733c-4c80-83c3-06665b85ade7", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-232939311-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931187d9051349cfaa9cfeeb2ddf032c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cf009f7-22", "ovs_interfaceid": "2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1123.823442] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:ca:50', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '225b6979-9329-403b-91fa-138bd41f6e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1123.831099] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Creating folder: Project (931187d9051349cfaa9cfeeb2ddf032c). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1123.831369] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-33550207-606c-4e8c-8a39-b4418a49f38d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.846006] env[61987]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1123.846173] env[61987]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61987) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1123.846504] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Folder already exists: Project (931187d9051349cfaa9cfeeb2ddf032c). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1123.846707] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Creating folder: Instances. Parent ref: group-v234438. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1123.846933] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c376b2fb-20c4-4aa9-b1aa-ad84851a0454 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.856435] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Created folder: Instances in parent group-v234438. [ 1123.856668] env[61987]: DEBUG oslo.service.loopingcall [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1123.856867] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1123.857112] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7337c504-61f1-42b5-939b-486b28aafbd6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.875941] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1123.875941] env[61987]: value = "task-1062380" [ 1123.875941] env[61987]: _type = "Task" [ 1123.875941] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.884683] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062380, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.063137] env[61987]: DEBUG nova.compute.manager [req-b6eced39-bb5c-4cad-a84d-5a74e9d4b6a4 req-768215de-55c0-4427-aaa3-ae0ebb1f51f4 service nova] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Received event network-changed-2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1124.063296] env[61987]: DEBUG nova.compute.manager [req-b6eced39-bb5c-4cad-a84d-5a74e9d4b6a4 req-768215de-55c0-4427-aaa3-ae0ebb1f51f4 service nova] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Refreshing instance network info cache due to event network-changed-2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1124.063526] env[61987]: DEBUG oslo_concurrency.lockutils [req-b6eced39-bb5c-4cad-a84d-5a74e9d4b6a4 req-768215de-55c0-4427-aaa3-ae0ebb1f51f4 service nova] Acquiring lock "refresh_cache-2ba77d32-d777-448e-8fa6-04920950f0f6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.063695] env[61987]: DEBUG oslo_concurrency.lockutils [req-b6eced39-bb5c-4cad-a84d-5a74e9d4b6a4 req-768215de-55c0-4427-aaa3-ae0ebb1f51f4 service nova] Acquired lock "refresh_cache-2ba77d32-d777-448e-8fa6-04920950f0f6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.063891] env[61987]: DEBUG nova.network.neutron [req-b6eced39-bb5c-4cad-a84d-5a74e9d4b6a4 req-768215de-55c0-4427-aaa3-ae0ebb1f51f4 service nova] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Refreshing network info cache for port 2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1124.268843] env[61987]: DEBUG nova.virt.hardware [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1124.269225] env[61987]: DEBUG nova.virt.hardware [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1124.269302] env[61987]: DEBUG nova.virt.hardware [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1124.269525] env[61987]: DEBUG nova.virt.hardware [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1124.269749] env[61987]: DEBUG nova.virt.hardware [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1124.269946] env[61987]: DEBUG nova.virt.hardware [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1124.270182] env[61987]: DEBUG nova.virt.hardware [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1124.270357] env[61987]: DEBUG nova.virt.hardware [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1124.270535] env[61987]: DEBUG nova.virt.hardware [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1124.270703] env[61987]: DEBUG nova.virt.hardware [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1124.270883] env[61987]: DEBUG nova.virt.hardware [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1124.271763] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73464b60-4680-4dd7-ae40-6f31163766f6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.279769] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c50a9afa-5683-4d75-9de3-37463c64758a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.292973] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:06:b5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'be8bd197-4b2b-46e7-88ea-2554b0438584', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9f0f3f6e-5109-4c21-9c31-5636c1f60bbc', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1124.300280] env[61987]: DEBUG oslo.service.loopingcall [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1124.300521] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1124.300729] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6195d590-8d89-4341-9eb4-0653d55ef6ec {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.321247] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1124.321247] env[61987]: value = "task-1062381" [ 1124.321247] env[61987]: _type = "Task" [ 1124.321247] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.329710] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062381, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.386835] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062380, 'name': CreateVM_Task, 'duration_secs': 0.300247} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.387054] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1124.387843] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'device_type': None, 'attachment_id': '79215acc-dd55-4305-8076-ecbf5a58971a', 'mount_device': '/dev/sda', 'delete_on_termination': True, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234441', 'volume_id': '4ab1a236-de8d-4578-8274-89bc106faa2e', 'name': 'volume-4ab1a236-de8d-4578-8274-89bc106faa2e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2ba77d32-d777-448e-8fa6-04920950f0f6', 'attached_at': '', 'detached_at': '', 'volume_id': '4ab1a236-de8d-4578-8274-89bc106faa2e', 'serial': '4ab1a236-de8d-4578-8274-89bc106faa2e'}, 'boot_index': 0, 'guest_format': None, 'disk_bus': None, 'volume_type': None}], 'swap': None} {{(pid=61987) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1124.388147] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Root volume attach. Driver type: vmdk {{(pid=61987) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1124.388969] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc681958-0993-47cd-824f-bbc7eeed97b4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.396858] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3555808-6bc3-4da7-a81b-734a528ecffd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.403069] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40f35811-073d-4398-a3b2-874b241b0966 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.408706] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-d06a408c-07ea-4813-9e6b-ebd4cf5e81c0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.415454] env[61987]: DEBUG oslo_vmware.api [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Waiting for the task: (returnval){ [ 1124.415454] env[61987]: value = "task-1062382" [ 1124.415454] env[61987]: _type = "Task" [ 1124.415454] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.423538] env[61987]: DEBUG oslo_vmware.api [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062382, 'name': RelocateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.831142] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062381, 'name': CreateVM_Task, 'duration_secs': 0.288109} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.831322] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1124.832000] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.832197] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.832546] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1124.832895] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03b26674-971b-487c-bd4c-ed9e0eccbe02 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.837698] env[61987]: DEBUG nova.network.neutron [req-b6eced39-bb5c-4cad-a84d-5a74e9d4b6a4 req-768215de-55c0-4427-aaa3-ae0ebb1f51f4 service nova] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Updated VIF entry in instance network info cache for port 2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1124.838397] env[61987]: DEBUG nova.network.neutron [req-b6eced39-bb5c-4cad-a84d-5a74e9d4b6a4 req-768215de-55c0-4427-aaa3-ae0ebb1f51f4 service nova] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Updating instance_info_cache with network_info: [{"id": "2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a", "address": "fa:16:3e:99:ca:50", "network": {"id": "b42b6e20-733c-4c80-83c3-06665b85ade7", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-232939311-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931187d9051349cfaa9cfeeb2ddf032c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cf009f7-22", "ovs_interfaceid": "2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.840291] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1124.840291] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52cd71ab-5dce-cd88-db61-393f369301b3" [ 1124.840291] env[61987]: _type = "Task" [ 1124.840291] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.848990] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52cd71ab-5dce-cd88-db61-393f369301b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.926042] env[61987]: DEBUG oslo_vmware.api [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062382, 'name': RelocateVM_Task} progress is 43%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.342387] env[61987]: DEBUG oslo_concurrency.lockutils [req-b6eced39-bb5c-4cad-a84d-5a74e9d4b6a4 req-768215de-55c0-4427-aaa3-ae0ebb1f51f4 service nova] Releasing lock "refresh_cache-2ba77d32-d777-448e-8fa6-04920950f0f6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.354035] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52cd71ab-5dce-cd88-db61-393f369301b3, 'name': SearchDatastore_Task, 'duration_secs': 0.013714} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.354402] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.354647] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1125.354917] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.355130] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.355335] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1125.355613] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4c895a56-a961-4b93-a83e-97566989316d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.364510] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1125.364655] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1125.365788] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-376e885d-40f0-4268-9c82-b134c7a958b6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.371117] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1125.371117] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]520ef9e5-8045-dd9e-e3c8-10856ba76218" [ 1125.371117] env[61987]: _type = "Task" [ 1125.371117] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.378880] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]520ef9e5-8045-dd9e-e3c8-10856ba76218, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.426208] env[61987]: DEBUG oslo_vmware.api [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062382, 'name': RelocateVM_Task} progress is 60%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.881840] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]520ef9e5-8045-dd9e-e3c8-10856ba76218, 'name': SearchDatastore_Task, 'duration_secs': 0.008678} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.882729] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a57679ea-d078-40e0-9e3d-2bc861e1e277 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.888456] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1125.888456] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52dc5d29-4c6d-d21b-bb84-3e66c40e424d" [ 1125.888456] env[61987]: _type = "Task" [ 1125.888456] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.896275] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52dc5d29-4c6d-d21b-bb84-3e66c40e424d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.927322] env[61987]: DEBUG oslo_vmware.api [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062382, 'name': RelocateVM_Task} progress is 75%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.399624] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52dc5d29-4c6d-d21b-bb84-3e66c40e424d, 'name': SearchDatastore_Task, 'duration_secs': 0.009955} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.400021] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.400161] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 76707afe-29ae-4257-a237-ba5a47b2511b/76707afe-29ae-4257-a237-ba5a47b2511b.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1126.400449] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-07708d7b-b466-4248-9fe3-62d7c59b1e89 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.407435] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1126.407435] env[61987]: value = "task-1062383" [ 1126.407435] env[61987]: _type = "Task" [ 1126.407435] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.415674] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062383, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.427020] env[61987]: DEBUG oslo_vmware.api [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062382, 'name': RelocateVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.662825] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "f030044b-bc6b-478c-8f63-d519aa76203c" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.663258] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "f030044b-bc6b-478c-8f63-d519aa76203c" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.663585] env[61987]: INFO nova.compute.manager [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Shelving [ 1126.916822] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062383, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.451638} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.917132] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 76707afe-29ae-4257-a237-ba5a47b2511b/76707afe-29ae-4257-a237-ba5a47b2511b.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1126.917392] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1126.917717] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5ae657b8-11c1-4787-be7f-46175e8e14c4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.928267] env[61987]: DEBUG oslo_vmware.api [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062382, 'name': RelocateVM_Task} progress is 97%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.929322] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1126.929322] env[61987]: value = "task-1062384" [ 1126.929322] env[61987]: _type = "Task" [ 1126.929322] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.936667] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062384, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.430616] env[61987]: DEBUG oslo_vmware.api [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062382, 'name': RelocateVM_Task, 'duration_secs': 2.926624} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.433822] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Volume attach. Driver type: vmdk {{(pid=61987) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1127.434054] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234441', 'volume_id': '4ab1a236-de8d-4578-8274-89bc106faa2e', 'name': 'volume-4ab1a236-de8d-4578-8274-89bc106faa2e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2ba77d32-d777-448e-8fa6-04920950f0f6', 'attached_at': '', 'detached_at': '', 'volume_id': '4ab1a236-de8d-4578-8274-89bc106faa2e', 'serial': '4ab1a236-de8d-4578-8274-89bc106faa2e'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1127.434818] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-670bcc62-fe95-4481-b281-907448fbb7c9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.441628] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062384, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069149} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.451769] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1127.452624] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f54b2a-6407-41ee-911e-a8104395100f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.455254] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54a3c540-514c-4c39-adf1-6dbd428fd64f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.476675] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Reconfiguring VM instance instance-00000073 to attach disk [datastore1] volume-4ab1a236-de8d-4578-8274-89bc106faa2e/volume-4ab1a236-de8d-4578-8274-89bc106faa2e.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1127.485283] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-321d8046-e60a-4a19-81c5-e112871c9bf4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.507504] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] 76707afe-29ae-4257-a237-ba5a47b2511b/76707afe-29ae-4257-a237-ba5a47b2511b.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1127.507751] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a6bca1f-8f0f-4e9a-9fd6-05fa67825561 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.526384] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1127.526384] env[61987]: value = "task-1062386" [ 1127.526384] env[61987]: _type = "Task" [ 1127.526384] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.527562] env[61987]: DEBUG oslo_vmware.api [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Waiting for the task: (returnval){ [ 1127.527562] env[61987]: value = "task-1062385" [ 1127.527562] env[61987]: _type = "Task" [ 1127.527562] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.538536] env[61987]: DEBUG oslo_vmware.api [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062385, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.541356] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062386, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.674331] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1127.674668] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5b847167-678b-4c07-8bb2-dcbad27f81b4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.681513] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1127.681513] env[61987]: value = "task-1062387" [ 1127.681513] env[61987]: _type = "Task" [ 1127.681513] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.689359] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062387, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.039277] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062386, 'name': ReconfigVM_Task, 'duration_secs': 0.362831} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.042334] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Reconfigured VM instance instance-00000070 to attach disk [datastore1] 76707afe-29ae-4257-a237-ba5a47b2511b/76707afe-29ae-4257-a237-ba5a47b2511b.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1128.043690] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'device_type': 'disk', 'device_name': '/dev/sda', 'encryption_secret_uuid': None, 'encryption_format': None, 'encryption_options': None, 'boot_index': 0, 'guest_format': None, 'size': 0, 'encrypted': False, 'disk_bus': None, 'image_id': '7bf8969f-d179-42af-9e15-527d5449f929'}], 'ephemerals': [], 'block_device_mapping': [{'device_type': None, 'attachment_id': '7f5bb1a0-8feb-4d80-9162-5b52924d8439', 'mount_device': '/dev/sdb', 'delete_on_termination': False, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234443', 'volume_id': '0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'name': 'volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '76707afe-29ae-4257-a237-ba5a47b2511b', 'attached_at': '', 'detached_at': '', 'volume_id': '0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'serial': '0f768543-5a0b-4e1d-887c-6cb5f60afe63'}, 'boot_index': None, 'guest_format': None, 'disk_bus': None, 'volume_type': None}], 'swap': None} {{(pid=61987) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1128.043912] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Volume attach. Driver type: vmdk {{(pid=61987) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1128.044123] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234443', 'volume_id': '0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'name': 'volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '76707afe-29ae-4257-a237-ba5a47b2511b', 'attached_at': '', 'detached_at': '', 'volume_id': '0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'serial': '0f768543-5a0b-4e1d-887c-6cb5f60afe63'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1128.044428] env[61987]: DEBUG oslo_vmware.api [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062385, 'name': ReconfigVM_Task, 'duration_secs': 0.329199} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.045141] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8e1ed5-23ca-471c-9d03-cd5af6de3b8d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.047553] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Reconfigured VM instance instance-00000073 to attach disk [datastore1] volume-4ab1a236-de8d-4578-8274-89bc106faa2e/volume-4ab1a236-de8d-4578-8274-89bc106faa2e.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1128.052190] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a1ac2d8-fb44-4b50-aa94-a636f89b73be {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.065860] env[61987]: DEBUG oslo_vmware.api [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Waiting for the task: (returnval){ [ 1128.065860] env[61987]: value = "task-1062388" [ 1128.065860] env[61987]: _type = "Task" [ 1128.065860] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.079247] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047c91b5-35a5-4958-b5b2-675cdd42627b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.088042] env[61987]: DEBUG oslo_vmware.api [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062388, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.107615] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63/volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1128.107913] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e889832-88e3-41f3-bf37-e37aeb2765d4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.126037] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1128.126037] env[61987]: value = "task-1062389" [ 1128.126037] env[61987]: _type = "Task" [ 1128.126037] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.133398] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062389, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.191271] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062387, 'name': PowerOffVM_Task, 'duration_secs': 0.215235} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.191600] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1128.192431] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc17b615-fbb5-46fd-8cc5-7f5d81525e47 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.211462] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90613524-3a2e-4951-b67c-59b6092ca753 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.586341] env[61987]: DEBUG oslo_vmware.api [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062388, 'name': ReconfigVM_Task, 'duration_secs': 0.128892} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.586663] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234441', 'volume_id': '4ab1a236-de8d-4578-8274-89bc106faa2e', 'name': 'volume-4ab1a236-de8d-4578-8274-89bc106faa2e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2ba77d32-d777-448e-8fa6-04920950f0f6', 'attached_at': '', 'detached_at': '', 'volume_id': '4ab1a236-de8d-4578-8274-89bc106faa2e', 'serial': '4ab1a236-de8d-4578-8274-89bc106faa2e'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1128.587181] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7af21a71-f0ab-479b-8352-16f9dd0eef2c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.593484] env[61987]: DEBUG oslo_vmware.api [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Waiting for the task: (returnval){ [ 1128.593484] env[61987]: value = "task-1062390" [ 1128.593484] env[61987]: _type = "Task" [ 1128.593484] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.601877] env[61987]: DEBUG oslo_vmware.api [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062390, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.634213] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062389, 'name': ReconfigVM_Task, 'duration_secs': 0.295826} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.634491] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Reconfigured VM instance instance-00000070 to attach disk [datastore2] volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63/volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1128.639170] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27526df4-991d-4f5b-9363-f44c874a64cf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.653750] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1128.653750] env[61987]: value = "task-1062391" [ 1128.653750] env[61987]: _type = "Task" [ 1128.653750] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.661411] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062391, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.723246] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Creating Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1128.723576] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b4a4c0fa-1466-4904-b839-2f9fb05fd7e9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.730457] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1128.730457] env[61987]: value = "task-1062392" [ 1128.730457] env[61987]: _type = "Task" [ 1128.730457] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.738738] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062392, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.103961] env[61987]: DEBUG oslo_vmware.api [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062390, 'name': Rename_Task, 'duration_secs': 0.171344} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.104299] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1129.104551] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6e9b07b7-dc7e-4cb2-875a-01f088ff0cd6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.111405] env[61987]: DEBUG oslo_vmware.api [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Waiting for the task: (returnval){ [ 1129.111405] env[61987]: value = "task-1062393" [ 1129.111405] env[61987]: _type = "Task" [ 1129.111405] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.119598] env[61987]: DEBUG oslo_vmware.api [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062393, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.166051] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062391, 'name': ReconfigVM_Task, 'duration_secs': 0.158409} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.166405] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234443', 'volume_id': '0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'name': 'volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '76707afe-29ae-4257-a237-ba5a47b2511b', 'attached_at': '', 'detached_at': '', 'volume_id': '0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'serial': '0f768543-5a0b-4e1d-887c-6cb5f60afe63'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1129.167087] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1c422ab5-6ef9-4f42-ba52-f4df8018e540 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.173656] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1129.173656] env[61987]: value = "task-1062394" [ 1129.173656] env[61987]: _type = "Task" [ 1129.173656] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.182836] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062394, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.240563] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062392, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.621064] env[61987]: DEBUG oslo_vmware.api [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062393, 'name': PowerOnVM_Task, 'duration_secs': 0.479577} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.621550] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1129.621550] env[61987]: INFO nova.compute.manager [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Took 6.08 seconds to spawn the instance on the hypervisor. [ 1129.621774] env[61987]: DEBUG nova.compute.manager [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1129.622507] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67aaa7eb-43d9-456d-a8a3-967544ab8af7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.683728] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062394, 'name': Rename_Task, 'duration_secs': 0.158244} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.684047] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1129.684311] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5823c006-45b2-4e3f-81f3-0431e14f7546 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.690445] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1129.690445] env[61987]: value = "task-1062395" [ 1129.690445] env[61987]: _type = "Task" [ 1129.690445] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.697714] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062395, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.740343] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062392, 'name': CreateSnapshot_Task, 'duration_secs': 0.51047} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.740629] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Created Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1129.741370] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62294ea0-966b-4d1d-9f81-d7e3209cffbc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.139940] env[61987]: INFO nova.compute.manager [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Took 12.52 seconds to build instance. [ 1130.200941] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062395, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.259066] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Creating linked-clone VM from snapshot {{(pid=61987) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1130.259416] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b7995315-9c70-49e6-8c44-d1b870dc85f3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.268073] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1130.268073] env[61987]: value = "task-1062396" [ 1130.268073] env[61987]: _type = "Task" [ 1130.268073] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.276971] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062396, 'name': CloneVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.447968] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ea184c7d-9a9b-41ea-ac15-b95549338848 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "7002962c-76ed-4a8d-be2a-b2b118847fea" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.448878] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ea184c7d-9a9b-41ea-ac15-b95549338848 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "7002962c-76ed-4a8d-be2a-b2b118847fea" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.641755] env[61987]: DEBUG oslo_concurrency.lockutils [None req-799d00d1-01a4-4d6f-aa38-591cccf598cf tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Lock "2ba77d32-d777-448e-8fa6-04920950f0f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.031s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.701671] env[61987]: DEBUG oslo_vmware.api [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062395, 'name': PowerOnVM_Task, 'duration_secs': 0.56358} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.701986] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1130.702234] env[61987]: DEBUG nova.compute.manager [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1130.703074] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd1593f6-0203-4650-b0cb-9efc3413600d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.779468] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062396, 'name': CloneVM_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.951666] env[61987]: INFO nova.compute.manager [None req-ea184c7d-9a9b-41ea-ac15-b95549338848 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Detaching volume dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb [ 1130.988994] env[61987]: INFO nova.virt.block_device [None req-ea184c7d-9a9b-41ea-ac15-b95549338848 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Attempting to driver detach volume dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb from mountpoint /dev/sdb [ 1130.989475] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea184c7d-9a9b-41ea-ac15-b95549338848 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Volume detach. Driver type: vmdk {{(pid=61987) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1130.989475] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea184c7d-9a9b-41ea-ac15-b95549338848 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234432', 'volume_id': 'dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb', 'name': 'volume-dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7002962c-76ed-4a8d-be2a-b2b118847fea', 'attached_at': '', 'detached_at': '', 'volume_id': 'dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb', 'serial': 'dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1130.990418] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec00155e-cc96-408b-ae72-a5923c8a4477 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.014033] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a38957-f0d3-41e0-bb2a-15f6f39d9dfa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.021396] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e03becd0-86a6-4b65-a010-b2a1839379bc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.043126] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c5c8d8e-8464-4fc3-b5c4-e0b3032622c1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.062053] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea184c7d-9a9b-41ea-ac15-b95549338848 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] The volume has not been displaced from its original location: [datastore1] volume-dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb/volume-dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb.vmdk. No consolidation needed. {{(pid=61987) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1131.067507] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea184c7d-9a9b-41ea-ac15-b95549338848 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Reconfiguring VM instance instance-0000006e to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1131.070239] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c650037a-2ecf-4ee5-9b2b-79c92b368714 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.088978] env[61987]: DEBUG nova.compute.manager [req-d255e1d2-4156-4502-aec6-585daf09ff31 req-91b647e9-8029-4324-80df-097d68119fc0 service nova] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Received event network-changed-2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1131.089226] env[61987]: DEBUG nova.compute.manager [req-d255e1d2-4156-4502-aec6-585daf09ff31 req-91b647e9-8029-4324-80df-097d68119fc0 service nova] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Refreshing instance network info cache due to event network-changed-2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1131.089473] env[61987]: DEBUG oslo_concurrency.lockutils [req-d255e1d2-4156-4502-aec6-585daf09ff31 req-91b647e9-8029-4324-80df-097d68119fc0 service nova] Acquiring lock "refresh_cache-2ba77d32-d777-448e-8fa6-04920950f0f6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.089617] env[61987]: DEBUG oslo_concurrency.lockutils [req-d255e1d2-4156-4502-aec6-585daf09ff31 req-91b647e9-8029-4324-80df-097d68119fc0 service nova] Acquired lock "refresh_cache-2ba77d32-d777-448e-8fa6-04920950f0f6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.089821] env[61987]: DEBUG nova.network.neutron [req-d255e1d2-4156-4502-aec6-585daf09ff31 req-91b647e9-8029-4324-80df-097d68119fc0 service nova] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Refreshing network info cache for port 2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1131.098480] env[61987]: DEBUG oslo_vmware.api [None req-ea184c7d-9a9b-41ea-ac15-b95549338848 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1131.098480] env[61987]: value = "task-1062397" [ 1131.098480] env[61987]: _type = "Task" [ 1131.098480] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.109646] env[61987]: DEBUG oslo_vmware.api [None req-ea184c7d-9a9b-41ea-ac15-b95549338848 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062397, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.221604] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.221604] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.221604] env[61987]: DEBUG nova.objects.instance [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61987) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1131.279378] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062396, 'name': CloneVM_Task, 'duration_secs': 0.977638} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.279729] env[61987]: INFO nova.virt.vmwareapi.vmops [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Created linked-clone VM from snapshot [ 1131.280551] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a2883d1-c9c9-4e73-bc83-d2accd9cfc5b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.288233] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Uploading image 9428b701-5c62-47e3-b19a-c8b536d18204 {{(pid=61987) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1131.312480] env[61987]: DEBUG oslo_vmware.rw_handles [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1131.312480] env[61987]: value = "vm-234448" [ 1131.312480] env[61987]: _type = "VirtualMachine" [ 1131.312480] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1131.312765] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8330bc66-77a6-479a-8369-bd5c454aea51 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.318803] env[61987]: DEBUG oslo_vmware.rw_handles [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lease: (returnval){ [ 1131.318803] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]528e5eb9-5491-5132-b45e-2d8868fe08af" [ 1131.318803] env[61987]: _type = "HttpNfcLease" [ 1131.318803] env[61987]: } obtained for exporting VM: (result){ [ 1131.318803] env[61987]: value = "vm-234448" [ 1131.318803] env[61987]: _type = "VirtualMachine" [ 1131.318803] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1131.319087] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the lease: (returnval){ [ 1131.319087] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]528e5eb9-5491-5132-b45e-2d8868fe08af" [ 1131.319087] env[61987]: _type = "HttpNfcLease" [ 1131.319087] env[61987]: } to be ready. {{(pid=61987) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1131.325482] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1131.325482] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]528e5eb9-5491-5132-b45e-2d8868fe08af" [ 1131.325482] env[61987]: _type = "HttpNfcLease" [ 1131.325482] env[61987]: } is initializing. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1131.598519] env[61987]: DEBUG nova.compute.manager [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Stashing vm_state: active {{(pid=61987) _prep_resize /opt/stack/nova/nova/compute/manager.py:6090}} [ 1131.613812] env[61987]: DEBUG oslo_vmware.api [None req-ea184c7d-9a9b-41ea-ac15-b95549338848 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062397, 'name': ReconfigVM_Task, 'duration_secs': 0.234655} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.614146] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea184c7d-9a9b-41ea-ac15-b95549338848 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Reconfigured VM instance instance-0000006e to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1131.619086] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3693f10b-8580-4793-85d3-23e020f9a188 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.636991] env[61987]: DEBUG oslo_vmware.api [None req-ea184c7d-9a9b-41ea-ac15-b95549338848 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1131.636991] env[61987]: value = "task-1062399" [ 1131.636991] env[61987]: _type = "Task" [ 1131.636991] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.646272] env[61987]: DEBUG oslo_vmware.api [None req-ea184c7d-9a9b-41ea-ac15-b95549338848 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062399, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.827845] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1131.827845] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]528e5eb9-5491-5132-b45e-2d8868fe08af" [ 1131.827845] env[61987]: _type = "HttpNfcLease" [ 1131.827845] env[61987]: } is ready. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1131.828232] env[61987]: DEBUG oslo_vmware.rw_handles [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1131.828232] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]528e5eb9-5491-5132-b45e-2d8868fe08af" [ 1131.828232] env[61987]: _type = "HttpNfcLease" [ 1131.828232] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1131.829057] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-565c2d8d-0f7b-40a0-a435-c274778a5d49 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.836713] env[61987]: DEBUG oslo_vmware.rw_handles [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529a4814-84fa-6732-d65f-444f782179fc/disk-0.vmdk from lease info. {{(pid=61987) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1131.836894] env[61987]: DEBUG oslo_vmware.rw_handles [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529a4814-84fa-6732-d65f-444f782179fc/disk-0.vmdk for reading. {{(pid=61987) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1132.026640] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-abef8479-0936-4654-a9bb-cf146a972e82 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.097374] env[61987]: DEBUG nova.network.neutron [req-d255e1d2-4156-4502-aec6-585daf09ff31 req-91b647e9-8029-4324-80df-097d68119fc0 service nova] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Updated VIF entry in instance network info cache for port 2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1132.097816] env[61987]: DEBUG nova.network.neutron [req-d255e1d2-4156-4502-aec6-585daf09ff31 req-91b647e9-8029-4324-80df-097d68119fc0 service nova] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Updating instance_info_cache with network_info: [{"id": "2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a", "address": "fa:16:3e:99:ca:50", "network": {"id": "b42b6e20-733c-4c80-83c3-06665b85ade7", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-232939311-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.246", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "931187d9051349cfaa9cfeeb2ddf032c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cf009f7-22", "ovs_interfaceid": "2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.120736] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.148091] env[61987]: DEBUG oslo_vmware.api [None req-ea184c7d-9a9b-41ea-ac15-b95549338848 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062399, 'name': ReconfigVM_Task, 'duration_secs': 0.137654} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.148091] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea184c7d-9a9b-41ea-ac15-b95549338848 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234432', 'volume_id': 'dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb', 'name': 'volume-dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7002962c-76ed-4a8d-be2a-b2b118847fea', 'attached_at': '', 'detached_at': '', 'volume_id': 'dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb', 'serial': 'dbd9ef53-276a-42ce-9ec9-68ef9ebf3aeb'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1132.232059] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b7098f12-96d4-4b14-905e-25c312ceb615 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.233778] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.113s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.601682] env[61987]: DEBUG oslo_concurrency.lockutils [req-d255e1d2-4156-4502-aec6-585daf09ff31 req-91b647e9-8029-4324-80df-097d68119fc0 service nova] Releasing lock "refresh_cache-2ba77d32-d777-448e-8fa6-04920950f0f6" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.684619] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6e2be32d-4e17-487e-aeb2-18be318168f6 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "76707afe-29ae-4257-a237-ba5a47b2511b" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.684994] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6e2be32d-4e17-487e-aeb2-18be318168f6 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "76707afe-29ae-4257-a237-ba5a47b2511b" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.692838] env[61987]: DEBUG nova.objects.instance [None req-ea184c7d-9a9b-41ea-ac15-b95549338848 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lazy-loading 'flavor' on Instance uuid 7002962c-76ed-4a8d-be2a-b2b118847fea {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1132.739388] env[61987]: INFO nova.compute.claims [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1133.188386] env[61987]: INFO nova.compute.manager [None req-6e2be32d-4e17-487e-aeb2-18be318168f6 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Detaching volume 0f768543-5a0b-4e1d-887c-6cb5f60afe63 [ 1133.228037] env[61987]: INFO nova.virt.block_device [None req-6e2be32d-4e17-487e-aeb2-18be318168f6 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Attempting to driver detach volume 0f768543-5a0b-4e1d-887c-6cb5f60afe63 from mountpoint /dev/sdb [ 1133.228037] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e2be32d-4e17-487e-aeb2-18be318168f6 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Volume detach. Driver type: vmdk {{(pid=61987) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1133.228037] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e2be32d-4e17-487e-aeb2-18be318168f6 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234443', 'volume_id': '0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'name': 'volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '76707afe-29ae-4257-a237-ba5a47b2511b', 'attached_at': '', 'detached_at': '', 'volume_id': '0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'serial': '0f768543-5a0b-4e1d-887c-6cb5f60afe63'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1133.229174] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e48b845-f1d5-41b3-971a-685e835e3b5a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.253256] env[61987]: INFO nova.compute.resource_tracker [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Updating resource usage from migration bf6d9417-391f-4b9b-8181-6d334abb9f70 [ 1133.256985] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a9b9b56-d0fa-4477-ad4b-23824781dbc0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.264529] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a56cee5a-b627-4988-ac32-3223c70f1b10 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.289118] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d5e6ba5-a5a6-487e-aff2-fb20b4aedcbb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.304390] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e2be32d-4e17-487e-aeb2-18be318168f6 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] The volume has not been displaced from its original location: [datastore2] volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63/volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63.vmdk. No consolidation needed. {{(pid=61987) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1133.309904] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e2be32d-4e17-487e-aeb2-18be318168f6 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Reconfiguring VM instance instance-00000070 to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1133.313076] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39efc367-2b10-4862-9bea-5a4c76426deb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.332425] env[61987]: DEBUG oslo_vmware.api [None req-6e2be32d-4e17-487e-aeb2-18be318168f6 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1133.332425] env[61987]: value = "task-1062400" [ 1133.332425] env[61987]: _type = "Task" [ 1133.332425] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.343611] env[61987]: DEBUG oslo_vmware.api [None req-6e2be32d-4e17-487e-aeb2-18be318168f6 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062400, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.417656] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16b2190d-6642-44dd-8cde-1b1579f5c802 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.425370] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f6e4aa-edf2-4b90-aab6-971ece108bd2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.456771] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6495fc44-daf9-49e9-b0d7-0853a022c8bc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.464435] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159794a2-fef8-4240-a65f-e235ed64f7e5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.479347] env[61987]: DEBUG nova.compute.provider_tree [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1133.703219] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ea184c7d-9a9b-41ea-ac15-b95549338848 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "7002962c-76ed-4a8d-be2a-b2b118847fea" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.255s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.842202] env[61987]: DEBUG oslo_vmware.api [None req-6e2be32d-4e17-487e-aeb2-18be318168f6 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062400, 'name': ReconfigVM_Task, 'duration_secs': 0.380571} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.842596] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e2be32d-4e17-487e-aeb2-18be318168f6 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Reconfigured VM instance instance-00000070 to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1133.847497] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a93fa6a-75d9-4cb0-b0fd-3a6112c7c3d5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.863308] env[61987]: DEBUG oslo_vmware.api [None req-6e2be32d-4e17-487e-aeb2-18be318168f6 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1133.863308] env[61987]: value = "task-1062401" [ 1133.863308] env[61987]: _type = "Task" [ 1133.863308] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.871532] env[61987]: DEBUG oslo_vmware.api [None req-6e2be32d-4e17-487e-aeb2-18be318168f6 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062401, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.982978] env[61987]: DEBUG nova.scheduler.client.report [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1134.375636] env[61987]: DEBUG oslo_vmware.api [None req-6e2be32d-4e17-487e-aeb2-18be318168f6 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062401, 'name': ReconfigVM_Task, 'duration_secs': 0.199638} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.375965] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e2be32d-4e17-487e-aeb2-18be318168f6 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234443', 'volume_id': '0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'name': 'volume-0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '76707afe-29ae-4257-a237-ba5a47b2511b', 'attached_at': '', 'detached_at': '', 'volume_id': '0f768543-5a0b-4e1d-887c-6cb5f60afe63', 'serial': '0f768543-5a0b-4e1d-887c-6cb5f60afe63'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1134.488380] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.255s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.488651] env[61987]: INFO nova.compute.manager [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Migrating [ 1134.769538] env[61987]: DEBUG oslo_concurrency.lockutils [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "7002962c-76ed-4a8d-be2a-b2b118847fea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.769952] env[61987]: DEBUG oslo_concurrency.lockutils [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "7002962c-76ed-4a8d-be2a-b2b118847fea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.770023] env[61987]: DEBUG oslo_concurrency.lockutils [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "7002962c-76ed-4a8d-be2a-b2b118847fea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.770215] env[61987]: DEBUG oslo_concurrency.lockutils [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "7002962c-76ed-4a8d-be2a-b2b118847fea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.770399] env[61987]: DEBUG oslo_concurrency.lockutils [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "7002962c-76ed-4a8d-be2a-b2b118847fea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.772723] env[61987]: INFO nova.compute.manager [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Terminating instance [ 1134.919784] env[61987]: DEBUG nova.objects.instance [None req-6e2be32d-4e17-487e-aeb2-18be318168f6 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lazy-loading 'flavor' on Instance uuid 76707afe-29ae-4257-a237-ba5a47b2511b {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1135.006843] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "refresh_cache-5912b687-2150-4133-a3a1-44ce38b3c3cb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1135.007156] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "refresh_cache-5912b687-2150-4133-a3a1-44ce38b3c3cb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.007527] env[61987]: DEBUG nova.network.neutron [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1135.277235] env[61987]: DEBUG nova.compute.manager [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1135.277532] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1135.278920] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca6ba5fe-50ed-4bb4-a6c2-6e42d76b56f5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.286921] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1135.287191] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-376d0ee5-8edc-48ea-bf47-68cf25508fef {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.293550] env[61987]: DEBUG oslo_vmware.api [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1135.293550] env[61987]: value = "task-1062402" [ 1135.293550] env[61987]: _type = "Task" [ 1135.293550] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.301586] env[61987]: DEBUG oslo_vmware.api [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062402, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.747372] env[61987]: DEBUG nova.network.neutron [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Updating instance_info_cache with network_info: [{"id": "a8217ebd-b52f-4664-a191-dc7f3513a366", "address": "fa:16:3e:49:1b:98", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8217ebd-b5", "ovs_interfaceid": "a8217ebd-b52f-4664-a191-dc7f3513a366", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.803626] env[61987]: DEBUG oslo_vmware.api [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062402, 'name': PowerOffVM_Task, 'duration_secs': 0.223892} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.803989] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1135.804182] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1135.804444] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bf8cbaf4-5ece-4a2c-81f7-0783674f30d8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.866071] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1135.866328] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1135.866511] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Deleting the datastore file [datastore1] 7002962c-76ed-4a8d-be2a-b2b118847fea {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1135.866794] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-444fa7a3-268c-463c-a2ab-7cf75f43e3a1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.872589] env[61987]: DEBUG oslo_vmware.api [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1135.872589] env[61987]: value = "task-1062404" [ 1135.872589] env[61987]: _type = "Task" [ 1135.872589] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.880719] env[61987]: DEBUG oslo_vmware.api [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062404, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.927932] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6e2be32d-4e17-487e-aeb2-18be318168f6 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "76707afe-29ae-4257-a237-ba5a47b2511b" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.243s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.250336] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "refresh_cache-5912b687-2150-4133-a3a1-44ce38b3c3cb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1136.382773] env[61987]: DEBUG oslo_vmware.api [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062404, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143282} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.383080] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1136.383281] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1136.383464] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1136.383650] env[61987]: INFO nova.compute.manager [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1136.383948] env[61987]: DEBUG oslo.service.loopingcall [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1136.384172] env[61987]: DEBUG nova.compute.manager [-] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1136.384268] env[61987]: DEBUG nova.network.neutron [-] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1136.492032] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1136.922540] env[61987]: DEBUG nova.compute.manager [req-3c60b58f-b0b4-4f77-b7c0-2865b9aaf12a req-8e7fe6f6-c88d-43c3-8e53-5102365befe5 service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Received event network-vif-deleted-d96dede6-c439-4a14-a4d1-62c44e03ea11 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1136.922908] env[61987]: INFO nova.compute.manager [req-3c60b58f-b0b4-4f77-b7c0-2865b9aaf12a req-8e7fe6f6-c88d-43c3-8e53-5102365befe5 service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Neutron deleted interface d96dede6-c439-4a14-a4d1-62c44e03ea11; detaching it from the instance and deleting it from the info cache [ 1136.922908] env[61987]: DEBUG nova.network.neutron [req-3c60b58f-b0b4-4f77-b7c0-2865b9aaf12a req-8e7fe6f6-c88d-43c3-8e53-5102365befe5 service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.945522] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "76707afe-29ae-4257-a237-ba5a47b2511b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.945972] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "76707afe-29ae-4257-a237-ba5a47b2511b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.946086] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "76707afe-29ae-4257-a237-ba5a47b2511b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.946238] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "76707afe-29ae-4257-a237-ba5a47b2511b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.946415] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "76707afe-29ae-4257-a237-ba5a47b2511b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.948574] env[61987]: INFO nova.compute.manager [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Terminating instance [ 1136.965516] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1136.965692] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61987) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11020}} [ 1136.966041] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1137.399293] env[61987]: DEBUG nova.network.neutron [-] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1137.426301] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7c4718e2-29b8-409c-98b7-b886d6b888fd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.435867] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-334da40c-860e-4a61-b59c-391fafdb9c4a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.451999] env[61987]: DEBUG nova.compute.manager [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1137.452303] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1137.453247] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f632438-114a-4200-9c9d-efb85e8170f8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.464074] env[61987]: DEBUG nova.compute.manager [req-3c60b58f-b0b4-4f77-b7c0-2865b9aaf12a req-8e7fe6f6-c88d-43c3-8e53-5102365befe5 service nova] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Detach interface failed, port_id=d96dede6-c439-4a14-a4d1-62c44e03ea11, reason: Instance 7002962c-76ed-4a8d-be2a-b2b118847fea could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 1137.470065] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.470065] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.470065] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.470065] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61987) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1137.470304] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1137.470883] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9cc19dd-944a-4431-bbf2-3bb72bcd01b1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.473483] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b4f7281b-d26a-4ebf-8b03-201a83b8daef {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.482434] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-515ce68d-1f1a-4f9e-b40e-8e6085312845 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.486473] env[61987]: DEBUG oslo_vmware.api [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1137.486473] env[61987]: value = "task-1062405" [ 1137.486473] env[61987]: _type = "Task" [ 1137.486473] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.499163] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630a75cd-1cb9-42c6-9402-6ecb82673597 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.505619] env[61987]: DEBUG oslo_vmware.api [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062405, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.510026] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb05cab5-7d5d-4f4d-ba81-36712ca77223 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.540503] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180098MB free_disk=179GB free_vcpus=48 pci_devices=None {{(pid=61987) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1137.540667] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.540873] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.767889] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05da146a-ce27-4699-b358-a3f47a5ae45c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.787667] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Updating instance '5912b687-2150-4133-a3a1-44ce38b3c3cb' progress to 0 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1137.901974] env[61987]: INFO nova.compute.manager [-] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Took 1.52 seconds to deallocate network for instance. [ 1138.000606] env[61987]: DEBUG oslo_vmware.api [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062405, 'name': PowerOffVM_Task, 'duration_secs': 0.218088} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.000985] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1138.000985] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1138.001863] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-daf8c612-5425-4368-b682-2933a3194801 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.070426] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1138.070658] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1138.070851] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Deleting the datastore file [datastore1] 76707afe-29ae-4257-a237-ba5a47b2511b {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1138.071139] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1ba765a8-4717-42c5-bd2b-11d8a2ca09d5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.078459] env[61987]: DEBUG oslo_vmware.api [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1138.078459] env[61987]: value = "task-1062407" [ 1138.078459] env[61987]: _type = "Task" [ 1138.078459] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.086134] env[61987]: DEBUG oslo_vmware.api [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062407, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.293694] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1138.294091] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb4aa84c-6fa9-43e8-985d-5e95ff2d5712 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.301070] env[61987]: DEBUG oslo_vmware.api [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1138.301070] env[61987]: value = "task-1062408" [ 1138.301070] env[61987]: _type = "Task" [ 1138.301070] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.316473] env[61987]: DEBUG oslo_vmware.api [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062408, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.408756] env[61987]: DEBUG oslo_concurrency.lockutils [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.551075] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Applying migration context for instance 5912b687-2150-4133-a3a1-44ce38b3c3cb as it has an incoming, in-progress migration bf6d9417-391f-4b9b-8181-6d334abb9f70. Migration status is migrating {{(pid=61987) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1138.551671] env[61987]: INFO nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Updating resource usage from migration bf6d9417-391f-4b9b-8181-6d334abb9f70 [ 1138.572178] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 4d15848d-5265-4897-bfa7-f3ad4c02ed83 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1138.572380] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1138.572554] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 473e0b73-6fcb-44c1-abf7-47d30c58af1b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1138.573180] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 7002962c-76ed-4a8d-be2a-b2b118847fea actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1138.573180] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 76707afe-29ae-4257-a237-ba5a47b2511b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1138.573180] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance f030044b-bc6b-478c-8f63-d519aa76203c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1138.573180] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 2ba77d32-d777-448e-8fa6-04920950f0f6 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1138.573456] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Migration bf6d9417-391f-4b9b-8181-6d334abb9f70 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1138.573456] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 5912b687-2150-4133-a3a1-44ce38b3c3cb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1138.573623] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1138.573794] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2304MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1138.589374] env[61987]: DEBUG oslo_vmware.api [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062407, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136132} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.592092] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1138.592343] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1138.592581] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1138.592790] env[61987]: INFO nova.compute.manager [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1138.593064] env[61987]: DEBUG oslo.service.loopingcall [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1138.593458] env[61987]: DEBUG nova.compute.manager [-] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1138.593561] env[61987]: DEBUG nova.network.neutron [-] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1138.695036] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1141362-1634-425b-99e9-2f5262438d52 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.702896] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae8018b7-cc8d-4d82-afb8-47a340da53b8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.733202] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59225c58-1397-47a4-9c55-75741b82cfe1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.740923] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-031e7c9b-644e-4e45-afd3-a8781667ed2a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.756397] env[61987]: DEBUG nova.compute.provider_tree [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1138.813326] env[61987]: DEBUG oslo_vmware.api [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062408, 'name': PowerOffVM_Task, 'duration_secs': 0.188215} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.813707] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1138.814067] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Updating instance '5912b687-2150-4133-a3a1-44ce38b3c3cb' progress to 17 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1139.036604] env[61987]: DEBUG nova.compute.manager [req-998f50d8-0142-44fd-b6a3-06bd270d8e53 req-e921b368-421c-4b93-bf74-f3ec1dd40d6d service nova] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Received event network-vif-deleted-9f0f3f6e-5109-4c21-9c31-5636c1f60bbc {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1139.036867] env[61987]: INFO nova.compute.manager [req-998f50d8-0142-44fd-b6a3-06bd270d8e53 req-e921b368-421c-4b93-bf74-f3ec1dd40d6d service nova] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Neutron deleted interface 9f0f3f6e-5109-4c21-9c31-5636c1f60bbc; detaching it from the instance and deleting it from the info cache [ 1139.036867] env[61987]: DEBUG nova.network.neutron [req-998f50d8-0142-44fd-b6a3-06bd270d8e53 req-e921b368-421c-4b93-bf74-f3ec1dd40d6d service nova] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.261909] env[61987]: DEBUG nova.scheduler.client.report [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1139.321165] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1139.321481] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1139.321672] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1139.321873] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1139.322038] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1139.322225] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1139.322488] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1139.322664] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1139.322864] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1139.323088] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1139.323284] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1139.328688] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c07a1d7-efe5-4a92-a81b-5d4f5191e4af {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.344549] env[61987]: DEBUG oslo_vmware.api [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1139.344549] env[61987]: value = "task-1062409" [ 1139.344549] env[61987]: _type = "Task" [ 1139.344549] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.353015] env[61987]: DEBUG oslo_vmware.api [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062409, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.518031] env[61987]: DEBUG nova.network.neutron [-] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.539558] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cb3983f5-f0d0-4e70-81af-8d875f9e60bc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.549585] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6c3727c-16cb-455e-8e98-ae18abf9131c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.576946] env[61987]: DEBUG nova.compute.manager [req-998f50d8-0142-44fd-b6a3-06bd270d8e53 req-e921b368-421c-4b93-bf74-f3ec1dd40d6d service nova] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Detach interface failed, port_id=9f0f3f6e-5109-4c21-9c31-5636c1f60bbc, reason: Instance 76707afe-29ae-4257-a237-ba5a47b2511b could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 1139.766193] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61987) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1139.766498] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.226s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.766847] env[61987]: DEBUG oslo_concurrency.lockutils [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.358s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.767140] env[61987]: DEBUG nova.objects.instance [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lazy-loading 'resources' on Instance uuid 7002962c-76ed-4a8d-be2a-b2b118847fea {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1139.855506] env[61987]: DEBUG oslo_vmware.api [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062409, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.020059] env[61987]: INFO nova.compute.manager [-] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Took 1.43 seconds to deallocate network for instance. [ 1140.360823] env[61987]: DEBUG oslo_vmware.api [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062409, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.400853] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b409c8d-8466-419e-ab51-2e0aacbe478b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.408555] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f37f8326-0bd4-4a51-aa3b-b207abc1ecca {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.438834] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a212c9-908e-49f8-a52a-766efd6ce1a1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.446116] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a90f291-b108-4d7f-9284-d6afbadf95af {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.459048] env[61987]: DEBUG nova.compute.provider_tree [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1140.527112] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.763021] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1140.763273] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1140.763424] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Starting heal instance info cache {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10401}} [ 1140.856153] env[61987]: DEBUG oslo_vmware.api [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062409, 'name': ReconfigVM_Task, 'duration_secs': 1.207166} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.856507] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Updating instance '5912b687-2150-4133-a3a1-44ce38b3c3cb' progress to 33 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1140.962804] env[61987]: DEBUG nova.scheduler.client.report [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1141.364024] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1141.364024] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1141.364024] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1141.364024] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1141.364896] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1141.365135] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1141.365372] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1141.365548] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1141.365728] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1141.365904] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1141.366098] env[61987]: DEBUG nova.virt.hardware [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1141.371522] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Reconfiguring VM instance instance-00000072 to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1141.371845] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82c49034-111c-46c2-ae30-8352498c8f23 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.391568] env[61987]: DEBUG oslo_vmware.api [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1141.391568] env[61987]: value = "task-1062410" [ 1141.391568] env[61987]: _type = "Task" [ 1141.391568] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.403095] env[61987]: DEBUG oslo_vmware.api [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062410, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.468147] env[61987]: DEBUG oslo_concurrency.lockutils [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.701s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.470610] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.944s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.470754] env[61987]: DEBUG nova.objects.instance [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lazy-loading 'resources' on Instance uuid 76707afe-29ae-4257-a237-ba5a47b2511b {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1141.490718] env[61987]: INFO nova.scheduler.client.report [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Deleted allocations for instance 7002962c-76ed-4a8d-be2a-b2b118847fea [ 1141.510341] env[61987]: DEBUG oslo_vmware.rw_handles [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529a4814-84fa-6732-d65f-444f782179fc/disk-0.vmdk. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1141.511291] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3805a63-e79e-4311-99c8-9f6bbf655dac {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.517662] env[61987]: DEBUG oslo_vmware.rw_handles [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529a4814-84fa-6732-d65f-444f782179fc/disk-0.vmdk is in state: ready. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1141.517889] env[61987]: ERROR oslo_vmware.rw_handles [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529a4814-84fa-6732-d65f-444f782179fc/disk-0.vmdk due to incomplete transfer. [ 1141.518133] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-2d4cd8c3-febf-4afd-9407-3f12a5adcbbd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.526872] env[61987]: DEBUG oslo_vmware.rw_handles [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529a4814-84fa-6732-d65f-444f782179fc/disk-0.vmdk. {{(pid=61987) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1141.527147] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Uploaded image 9428b701-5c62-47e3-b19a-c8b536d18204 to the Glance image server {{(pid=61987) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1141.529622] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Destroying the VM {{(pid=61987) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1141.530030] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-08857505-4693-4d1d-963d-ac8adf96ff02 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.536489] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1141.536489] env[61987]: value = "task-1062411" [ 1141.536489] env[61987]: _type = "Task" [ 1141.536489] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.544718] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062411, 'name': Destroy_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.769465] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Didn't find any instances for network info cache update. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10487}} [ 1141.769781] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1141.769973] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1141.770151] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1141.770303] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1141.900845] env[61987]: DEBUG oslo_vmware.api [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062410, 'name': ReconfigVM_Task, 'duration_secs': 0.162614} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.901132] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Reconfigured VM instance instance-00000072 to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1141.901888] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa63737c-68b9-416e-940e-82652b67aec4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.922938] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] 5912b687-2150-4133-a3a1-44ce38b3c3cb/5912b687-2150-4133-a3a1-44ce38b3c3cb.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1141.923148] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af84150f-62e9-4712-99ba-95d8f8dd3121 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.939503] env[61987]: DEBUG oslo_vmware.api [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1141.939503] env[61987]: value = "task-1062412" [ 1141.939503] env[61987]: _type = "Task" [ 1141.939503] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.946925] env[61987]: DEBUG oslo_vmware.api [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062412, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.000464] env[61987]: DEBUG oslo_concurrency.lockutils [None req-90dc4551-653b-44d8-afd8-7cd0da63e952 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "7002962c-76ed-4a8d-be2a-b2b118847fea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.231s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.048410] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062411, 'name': Destroy_Task, 'duration_secs': 0.334947} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.048410] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Destroyed the VM [ 1142.048691] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Deleting Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1142.048939] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-58e5b351-ddda-4c8e-8ce2-11f8fffd16c4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.055079] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1142.055079] env[61987]: value = "task-1062413" [ 1142.055079] env[61987]: _type = "Task" [ 1142.055079] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.064921] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062413, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.096267] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-281eba51-55ee-4ad7-8110-8ee7376b8873 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.104675] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1646099-9b9a-4016-806b-045f9df7e4d7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.137182] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65aef81f-0950-4965-91d3-7f0347ed6a15 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.144562] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12869980-9bd0-4591-8b3e-286ce1d58bed {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.158400] env[61987]: DEBUG nova.compute.provider_tree [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1142.449980] env[61987]: DEBUG oslo_vmware.api [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062412, 'name': ReconfigVM_Task, 'duration_secs': 0.297248} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.450403] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Reconfigured VM instance instance-00000072 to attach disk [datastore2] 5912b687-2150-4133-a3a1-44ce38b3c3cb/5912b687-2150-4133-a3a1-44ce38b3c3cb.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1142.450631] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Updating instance '5912b687-2150-4133-a3a1-44ce38b3c3cb' progress to 50 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1142.564656] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062413, 'name': RemoveSnapshot_Task, 'duration_secs': 0.362421} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.564945] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Deleted Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1142.565236] env[61987]: DEBUG nova.compute.manager [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1142.566015] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5138fa6-4736-4387-a92d-58da685b2b2f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.662009] env[61987]: DEBUG nova.scheduler.client.report [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1142.957424] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78fedf60-0054-4539-b193-3a8bcc5c87b1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.976082] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a246f145-fdac-4ccc-a4d9-f57b6146834e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.992541] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Updating instance '5912b687-2150-4133-a3a1-44ce38b3c3cb' progress to 67 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1143.076804] env[61987]: INFO nova.compute.manager [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Shelve offloading [ 1143.168026] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.697s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.189052] env[61987]: INFO nova.scheduler.client.report [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Deleted allocations for instance 76707afe-29ae-4257-a237-ba5a47b2511b [ 1143.533350] env[61987]: DEBUG nova.network.neutron [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Port a8217ebd-b52f-4664-a191-dc7f3513a366 binding to destination host cpu-1 is already ACTIVE {{(pid=61987) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1143.580518] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1143.580831] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec234a43-3f80-42e6-86ee-c82e373df122 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.588441] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1143.588441] env[61987]: value = "task-1062415" [ 1143.588441] env[61987]: _type = "Task" [ 1143.588441] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.598936] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] VM already powered off {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1143.599155] env[61987]: DEBUG nova.compute.manager [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1143.599866] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f032d1d-2bf8-41ea-a13e-50f1edabc007 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.605175] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "refresh_cache-f030044b-bc6b-478c-8f63-d519aa76203c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1143.605356] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquired lock "refresh_cache-f030044b-bc6b-478c-8f63-d519aa76203c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.605529] env[61987]: DEBUG nova.network.neutron [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1143.697026] env[61987]: DEBUG oslo_concurrency.lockutils [None req-cc73b6d5-a570-4345-8793-b1035d68ec05 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "76707afe-29ae-4257-a237-ba5a47b2511b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.751s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.423836] env[61987]: DEBUG nova.network.neutron [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Updating instance_info_cache with network_info: [{"id": "2d38bac7-bebb-462a-aac9-bbff9d711577", "address": "fa:16:3e:f8:d4:2d", "network": {"id": "eac55241-aad5-4ddb-a01e-625c9298a8af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1411727225-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90faa728219a432fa241b635071729e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f996252-e329-42bd-a897-446dfe2b81cd", "external-id": "nsx-vlan-transportzone-535", "segmentation_id": 535, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d38bac7-be", "ovs_interfaceid": "2d38bac7-bebb-462a-aac9-bbff9d711577", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.554962] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "5912b687-2150-4133-a3a1-44ce38b3c3cb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.555394] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "5912b687-2150-4133-a3a1-44ce38b3c3cb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.555394] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "5912b687-2150-4133-a3a1-44ce38b3c3cb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.928100] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Releasing lock "refresh_cache-f030044b-bc6b-478c-8f63-d519aa76203c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1145.241426] env[61987]: DEBUG nova.compute.manager [req-31ecfccf-6e33-43d6-a6a8-8323d837b1ae req-edd975a1-cf61-4ed0-9944-bd5c6b94128a service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Received event network-vif-unplugged-2d38bac7-bebb-462a-aac9-bbff9d711577 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1145.241663] env[61987]: DEBUG oslo_concurrency.lockutils [req-31ecfccf-6e33-43d6-a6a8-8323d837b1ae req-edd975a1-cf61-4ed0-9944-bd5c6b94128a service nova] Acquiring lock "f030044b-bc6b-478c-8f63-d519aa76203c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.241879] env[61987]: DEBUG oslo_concurrency.lockutils [req-31ecfccf-6e33-43d6-a6a8-8323d837b1ae req-edd975a1-cf61-4ed0-9944-bd5c6b94128a service nova] Lock "f030044b-bc6b-478c-8f63-d519aa76203c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.242138] env[61987]: DEBUG oslo_concurrency.lockutils [req-31ecfccf-6e33-43d6-a6a8-8323d837b1ae req-edd975a1-cf61-4ed0-9944-bd5c6b94128a service nova] Lock "f030044b-bc6b-478c-8f63-d519aa76203c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.242334] env[61987]: DEBUG nova.compute.manager [req-31ecfccf-6e33-43d6-a6a8-8323d837b1ae req-edd975a1-cf61-4ed0-9944-bd5c6b94128a service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] No waiting events found dispatching network-vif-unplugged-2d38bac7-bebb-462a-aac9-bbff9d711577 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1145.242516] env[61987]: WARNING nova.compute.manager [req-31ecfccf-6e33-43d6-a6a8-8323d837b1ae req-edd975a1-cf61-4ed0-9944-bd5c6b94128a service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Received unexpected event network-vif-unplugged-2d38bac7-bebb-462a-aac9-bbff9d711577 for instance with vm_state shelved and task_state shelving_offloading. [ 1145.335324] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1145.336249] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f4a048e-55f5-4531-a2fc-c52cb1494991 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.344144] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1145.344384] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1facbd48-8ddc-4024-9253-f928750fe80f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.403856] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1145.404039] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1145.405026] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Deleting the datastore file [datastore2] f030044b-bc6b-478c-8f63-d519aa76203c {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1145.405026] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-825c1d0e-36a2-42b6-9e1d-d3c5d49037d0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.411057] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1145.411057] env[61987]: value = "task-1062417" [ 1145.411057] env[61987]: _type = "Task" [ 1145.411057] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.419906] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062417, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.588858] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "refresh_cache-5912b687-2150-4133-a3a1-44ce38b3c3cb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1145.589130] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "refresh_cache-5912b687-2150-4133-a3a1-44ce38b3c3cb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.589312] env[61987]: DEBUG nova.network.neutron [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1145.923418] env[61987]: DEBUG oslo_vmware.api [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062417, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14073} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.923852] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1145.924136] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1145.924362] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1145.946655] env[61987]: INFO nova.scheduler.client.report [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Deleted allocations for instance f030044b-bc6b-478c-8f63-d519aa76203c [ 1146.010837] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.011170] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.351960] env[61987]: DEBUG nova.network.neutron [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Updating instance_info_cache with network_info: [{"id": "a8217ebd-b52f-4664-a191-dc7f3513a366", "address": "fa:16:3e:49:1b:98", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8217ebd-b5", "ovs_interfaceid": "a8217ebd-b52f-4664-a191-dc7f3513a366", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1146.451682] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.451973] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.452208] env[61987]: DEBUG nova.objects.instance [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lazy-loading 'resources' on Instance uuid f030044b-bc6b-478c-8f63-d519aa76203c {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1146.513601] env[61987]: DEBUG nova.compute.manager [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1146.854846] env[61987]: DEBUG oslo_concurrency.lockutils [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "refresh_cache-5912b687-2150-4133-a3a1-44ce38b3c3cb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.955089] env[61987]: DEBUG nova.objects.instance [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lazy-loading 'numa_topology' on Instance uuid f030044b-bc6b-478c-8f63-d519aa76203c {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1147.033662] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.104075] env[61987]: DEBUG oslo_concurrency.lockutils [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "473e0b73-6fcb-44c1-abf7-47d30c58af1b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.104379] env[61987]: DEBUG oslo_concurrency.lockutils [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "473e0b73-6fcb-44c1-abf7-47d30c58af1b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.104602] env[61987]: DEBUG oslo_concurrency.lockutils [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "473e0b73-6fcb-44c1-abf7-47d30c58af1b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.104798] env[61987]: DEBUG oslo_concurrency.lockutils [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "473e0b73-6fcb-44c1-abf7-47d30c58af1b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.104975] env[61987]: DEBUG oslo_concurrency.lockutils [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "473e0b73-6fcb-44c1-abf7-47d30c58af1b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.107247] env[61987]: INFO nova.compute.manager [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Terminating instance [ 1147.264571] env[61987]: DEBUG nova.compute.manager [req-c0a0a5fe-83e9-4324-b215-2143557cdbf1 req-db992649-19ef-44ce-ac13-41f62498247b service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Received event network-changed-2d38bac7-bebb-462a-aac9-bbff9d711577 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1147.264824] env[61987]: DEBUG nova.compute.manager [req-c0a0a5fe-83e9-4324-b215-2143557cdbf1 req-db992649-19ef-44ce-ac13-41f62498247b service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Refreshing instance network info cache due to event network-changed-2d38bac7-bebb-462a-aac9-bbff9d711577. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1147.264974] env[61987]: DEBUG oslo_concurrency.lockutils [req-c0a0a5fe-83e9-4324-b215-2143557cdbf1 req-db992649-19ef-44ce-ac13-41f62498247b service nova] Acquiring lock "refresh_cache-f030044b-bc6b-478c-8f63-d519aa76203c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1147.265141] env[61987]: DEBUG oslo_concurrency.lockutils [req-c0a0a5fe-83e9-4324-b215-2143557cdbf1 req-db992649-19ef-44ce-ac13-41f62498247b service nova] Acquired lock "refresh_cache-f030044b-bc6b-478c-8f63-d519aa76203c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.265310] env[61987]: DEBUG nova.network.neutron [req-c0a0a5fe-83e9-4324-b215-2143557cdbf1 req-db992649-19ef-44ce-ac13-41f62498247b service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Refreshing network info cache for port 2d38bac7-bebb-462a-aac9-bbff9d711577 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1147.374825] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d652cf7-4362-4309-b2de-c85ed54b288a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.394873] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a77ca2-944a-4402-8627-b8bac9fcb665 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.401946] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Updating instance '5912b687-2150-4133-a3a1-44ce38b3c3cb' progress to 83 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1147.457012] env[61987]: DEBUG nova.objects.base [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61987) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1147.551570] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fffe533c-e4dd-4008-86a5-61f731211221 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.559626] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f8130f2-71a2-436d-8a51-0aaa80878dfe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.593942] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4808feac-bb13-416b-83ac-cae2da893fdf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.602014] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95997510-0206-46a8-9ba6-b72ce20666df {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.617015] env[61987]: DEBUG nova.compute.manager [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1147.617241] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1147.617939] env[61987]: DEBUG nova.compute.provider_tree [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1147.619604] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2f0c487-9410-43f3-9a6e-a642ddcc77a5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.626286] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1147.626936] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5b81966f-b891-4de5-947a-d6992198deff {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.634275] env[61987]: DEBUG oslo_vmware.api [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1147.634275] env[61987]: value = "task-1062418" [ 1147.634275] env[61987]: _type = "Task" [ 1147.634275] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.641813] env[61987]: DEBUG oslo_vmware.api [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062418, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.907775] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1147.908090] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c080791b-137b-4af0-afae-d85110567a14 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.918150] env[61987]: DEBUG oslo_vmware.api [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1147.918150] env[61987]: value = "task-1062419" [ 1147.918150] env[61987]: _type = "Task" [ 1147.918150] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.928231] env[61987]: DEBUG oslo_vmware.api [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062419, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.994625] env[61987]: DEBUG nova.network.neutron [req-c0a0a5fe-83e9-4324-b215-2143557cdbf1 req-db992649-19ef-44ce-ac13-41f62498247b service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Updated VIF entry in instance network info cache for port 2d38bac7-bebb-462a-aac9-bbff9d711577. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1147.995064] env[61987]: DEBUG nova.network.neutron [req-c0a0a5fe-83e9-4324-b215-2143557cdbf1 req-db992649-19ef-44ce-ac13-41f62498247b service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Updating instance_info_cache with network_info: [{"id": "2d38bac7-bebb-462a-aac9-bbff9d711577", "address": "fa:16:3e:f8:d4:2d", "network": {"id": "eac55241-aad5-4ddb-a01e-625c9298a8af", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1411727225-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90faa728219a432fa241b635071729e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap2d38bac7-be", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.068064] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "f030044b-bc6b-478c-8f63-d519aa76203c" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1148.123378] env[61987]: DEBUG nova.scheduler.client.report [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1148.145867] env[61987]: DEBUG oslo_vmware.api [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062418, 'name': PowerOffVM_Task, 'duration_secs': 0.219694} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.146174] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1148.146365] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1148.146635] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7bd05068-a4b8-4aac-99d3-56d4dbbc6872 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.222149] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1148.222431] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1148.222621] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Deleting the datastore file [datastore2] 473e0b73-6fcb-44c1-abf7-47d30c58af1b {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1148.222914] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-14c4482d-0c09-4f5f-b3bf-818c2a843737 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.244819] env[61987]: DEBUG oslo_vmware.api [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1148.244819] env[61987]: value = "task-1062421" [ 1148.244819] env[61987]: _type = "Task" [ 1148.244819] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.254482] env[61987]: DEBUG oslo_vmware.api [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062421, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.428938] env[61987]: DEBUG oslo_vmware.api [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062419, 'name': PowerOnVM_Task, 'duration_secs': 0.370404} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.429255] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1148.429447] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-fe094151-da34-4160-9c40-a7c212890889 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Updating instance '5912b687-2150-4133-a3a1-44ce38b3c3cb' progress to 100 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1148.498056] env[61987]: DEBUG oslo_concurrency.lockutils [req-c0a0a5fe-83e9-4324-b215-2143557cdbf1 req-db992649-19ef-44ce-ac13-41f62498247b service nova] Releasing lock "refresh_cache-f030044b-bc6b-478c-8f63-d519aa76203c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1148.627950] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.176s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1148.630414] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.597s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1148.631820] env[61987]: INFO nova.compute.claims [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1148.754876] env[61987]: DEBUG oslo_vmware.api [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062421, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.200664} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.755207] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1148.755450] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1148.755694] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1148.755895] env[61987]: INFO nova.compute.manager [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1148.756158] env[61987]: DEBUG oslo.service.loopingcall [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1148.756362] env[61987]: DEBUG nova.compute.manager [-] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1148.756460] env[61987]: DEBUG nova.network.neutron [-] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1149.141295] env[61987]: DEBUG oslo_concurrency.lockutils [None req-c2a24fcb-6ef0-4af1-acad-5a138897950f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "f030044b-bc6b-478c-8f63-d519aa76203c" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 22.478s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1149.142154] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "f030044b-bc6b-478c-8f63-d519aa76203c" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.074s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1149.142355] env[61987]: INFO nova.compute.manager [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Unshelving [ 1149.291272] env[61987]: DEBUG nova.compute.manager [req-59e4f9f0-3e2c-4fc6-9f4d-47bce998b678 req-2ccf5688-7f4a-4823-a797-ebbf247827a6 service nova] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Received event network-vif-deleted-5e6a927f-3d25-42bc-86b7-39767ddfbd98 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1149.291481] env[61987]: INFO nova.compute.manager [req-59e4f9f0-3e2c-4fc6-9f4d-47bce998b678 req-2ccf5688-7f4a-4823-a797-ebbf247827a6 service nova] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Neutron deleted interface 5e6a927f-3d25-42bc-86b7-39767ddfbd98; detaching it from the instance and deleting it from the info cache [ 1149.291657] env[61987]: DEBUG nova.network.neutron [req-59e4f9f0-3e2c-4fc6-9f4d-47bce998b678 req-2ccf5688-7f4a-4823-a797-ebbf247827a6 service nova] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1149.453914] env[61987]: DEBUG nova.network.neutron [-] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1149.753839] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4161ce5a-914a-4e7b-b21b-2bc7c303eb2f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.762742] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de620294-4a2c-44fa-aed5-85c663fceb61 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.793701] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a48a792b-1c54-419b-b459-4960d6b0f878 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.796770] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bb55b9bd-80a3-429f-a0cc-66874eb883dd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.806030] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-067ef0a2-0fcb-466b-b543-f4c675113cc5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.812396] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a6efd72-5be0-406d-ba98-a1ac5434db95 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.835923] env[61987]: DEBUG nova.compute.provider_tree [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1149.844393] env[61987]: DEBUG nova.compute.manager [req-59e4f9f0-3e2c-4fc6-9f4d-47bce998b678 req-2ccf5688-7f4a-4823-a797-ebbf247827a6 service nova] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Detach interface failed, port_id=5e6a927f-3d25-42bc-86b7-39767ddfbd98, reason: Instance 473e0b73-6fcb-44c1-abf7-47d30c58af1b could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 1149.956262] env[61987]: INFO nova.compute.manager [-] [instance: 473e0b73-6fcb-44c1-abf7-47d30c58af1b] Took 1.20 seconds to deallocate network for instance. [ 1150.153710] env[61987]: DEBUG nova.compute.utils [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1150.319849] env[61987]: DEBUG nova.network.neutron [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Port a8217ebd-b52f-4664-a191-dc7f3513a366 binding to destination host cpu-1 is already ACTIVE {{(pid=61987) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1150.320040] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "refresh_cache-5912b687-2150-4133-a3a1-44ce38b3c3cb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1150.320213] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "refresh_cache-5912b687-2150-4133-a3a1-44ce38b3c3cb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.320387] env[61987]: DEBUG nova.network.neutron [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1150.339443] env[61987]: DEBUG nova.scheduler.client.report [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1150.463482] env[61987]: DEBUG oslo_concurrency.lockutils [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.657619] env[61987]: INFO nova.virt.block_device [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Booting with volume e6316fa6-4d06-40fb-90d8-60c4cfd7c454 at /dev/sdb [ 1150.703014] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-824f37e1-99e7-4ae1-9e2d-e956095e87f0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.713475] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a795e9fe-908e-4461-9200-fe33b9052638 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.740727] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f4b78bb5-8e72-4ef2-8d53-91e977a9ba0b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.749388] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51f1bce7-33f3-4b05-a91c-8b82c649bd49 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.777841] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac0cf986-5469-4d1d-aa26-c56b83402d7e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.784423] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfab9caa-1d97-4119-b2d3-8e4e9481ad36 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.798730] env[61987]: DEBUG nova.virt.block_device [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Updating existing volume attachment record: 8d4a2c70-5bb9-4c42-b08a-953ddb3121c4 {{(pid=61987) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1150.844280] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.214s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.844932] env[61987]: DEBUG nova.compute.manager [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1150.847613] env[61987]: DEBUG oslo_concurrency.lockutils [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.385s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.847864] env[61987]: DEBUG nova.objects.instance [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lazy-loading 'resources' on Instance uuid 473e0b73-6fcb-44c1-abf7-47d30c58af1b {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1151.035810] env[61987]: DEBUG nova.network.neutron [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Updating instance_info_cache with network_info: [{"id": "a8217ebd-b52f-4664-a191-dc7f3513a366", "address": "fa:16:3e:49:1b:98", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8217ebd-b5", "ovs_interfaceid": "a8217ebd-b52f-4664-a191-dc7f3513a366", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1151.350752] env[61987]: DEBUG nova.compute.utils [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1151.355748] env[61987]: DEBUG nova.compute.manager [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1151.355986] env[61987]: DEBUG nova.network.neutron [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1151.404851] env[61987]: DEBUG nova.policy [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c5cf7a7673cc4f5aadca7960c355c394', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ea5c65ac3b143178b13731663dbd8ee', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 1151.489021] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246c7c77-be1b-4028-8953-9cc9a54f71c7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.498167] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4c7a0c8-9a00-477d-8320-67103e382938 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.529135] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48ab3b0c-517f-443e-8a0e-6543ce62094c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.539019] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5dd7a5e-16e8-4838-a401-3dc9886b8026 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.542604] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "refresh_cache-5912b687-2150-4133-a3a1-44ce38b3c3cb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1151.554547] env[61987]: DEBUG nova.compute.provider_tree [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1151.649074] env[61987]: DEBUG nova.network.neutron [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Successfully created port: e48cc892-3cba-40a4-a803-6eab279a5d72 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1151.859784] env[61987]: DEBUG nova.compute.manager [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1152.046047] env[61987]: DEBUG nova.compute.manager [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61987) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:900}} [ 1152.046422] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.057606] env[61987]: DEBUG nova.scheduler.client.report [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1152.563053] env[61987]: DEBUG oslo_concurrency.lockutils [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.714s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.564541] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.518s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1152.584298] env[61987]: INFO nova.scheduler.client.report [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Deleted allocations for instance 473e0b73-6fcb-44c1-abf7-47d30c58af1b [ 1152.869127] env[61987]: DEBUG nova.compute.manager [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1152.893810] env[61987]: DEBUG nova.virt.hardware [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1152.894097] env[61987]: DEBUG nova.virt.hardware [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1152.894327] env[61987]: DEBUG nova.virt.hardware [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1152.894550] env[61987]: DEBUG nova.virt.hardware [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1152.894712] env[61987]: DEBUG nova.virt.hardware [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1152.894870] env[61987]: DEBUG nova.virt.hardware [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1152.895097] env[61987]: DEBUG nova.virt.hardware [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1152.895277] env[61987]: DEBUG nova.virt.hardware [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1152.895454] env[61987]: DEBUG nova.virt.hardware [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1152.895636] env[61987]: DEBUG nova.virt.hardware [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1152.895821] env[61987]: DEBUG nova.virt.hardware [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1152.896723] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b068092-a4de-4485-bfdc-a622deabe45c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.905662] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981d881f-6275-4f59-911c-79c73b4a1a05 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.070560] env[61987]: DEBUG nova.objects.instance [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lazy-loading 'migration_context' on Instance uuid 5912b687-2150-4133-a3a1-44ce38b3c3cb {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1153.092168] env[61987]: DEBUG oslo_concurrency.lockutils [None req-12f1f333-2576-46ab-8faf-da6853b052ad tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "473e0b73-6fcb-44c1-abf7-47d30c58af1b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.988s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.168613] env[61987]: DEBUG nova.compute.manager [req-dbe2e66f-6052-423c-8a9f-1ca03e62f586 req-684fc14c-85b1-49fe-818a-7d19c0fb6528 service nova] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Received event network-vif-plugged-e48cc892-3cba-40a4-a803-6eab279a5d72 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1153.168613] env[61987]: DEBUG oslo_concurrency.lockutils [req-dbe2e66f-6052-423c-8a9f-1ca03e62f586 req-684fc14c-85b1-49fe-818a-7d19c0fb6528 service nova] Acquiring lock "430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.168613] env[61987]: DEBUG oslo_concurrency.lockutils [req-dbe2e66f-6052-423c-8a9f-1ca03e62f586 req-684fc14c-85b1-49fe-818a-7d19c0fb6528 service nova] Lock "430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.168613] env[61987]: DEBUG oslo_concurrency.lockutils [req-dbe2e66f-6052-423c-8a9f-1ca03e62f586 req-684fc14c-85b1-49fe-818a-7d19c0fb6528 service nova] Lock "430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.169149] env[61987]: DEBUG nova.compute.manager [req-dbe2e66f-6052-423c-8a9f-1ca03e62f586 req-684fc14c-85b1-49fe-818a-7d19c0fb6528 service nova] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] No waiting events found dispatching network-vif-plugged-e48cc892-3cba-40a4-a803-6eab279a5d72 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1153.169149] env[61987]: WARNING nova.compute.manager [req-dbe2e66f-6052-423c-8a9f-1ca03e62f586 req-684fc14c-85b1-49fe-818a-7d19c0fb6528 service nova] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Received unexpected event network-vif-plugged-e48cc892-3cba-40a4-a803-6eab279a5d72 for instance with vm_state building and task_state spawning. [ 1153.265513] env[61987]: DEBUG nova.network.neutron [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Successfully updated port: e48cc892-3cba-40a4-a803-6eab279a5d72 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1153.658402] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa9e7641-e802-4f27-b4d0-adbcd86f5d21 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.666299] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-552a4f96-5aba-4d0f-bef9-beaafd31ea70 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.697293] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ee4fe4e-700b-4f8f-a60d-f1608e5e440e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.705658] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c848336-aa3c-4a2e-9518-d12853bfb174 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.719327] env[61987]: DEBUG nova.compute.provider_tree [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1153.761645] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.761917] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.762143] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.762342] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.762515] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.764524] env[61987]: INFO nova.compute.manager [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Terminating instance [ 1153.770735] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "refresh_cache-430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1153.770950] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired lock "refresh_cache-430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.772313] env[61987]: DEBUG nova.network.neutron [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1154.221988] env[61987]: DEBUG nova.scheduler.client.report [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1154.268037] env[61987]: DEBUG nova.compute.manager [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1154.268285] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1154.269307] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f3d52f-00ca-4a70-8be2-a441c9955684 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.277866] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1154.278386] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1f47c945-f6b3-4599-b3fb-8f5cf3f4443e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.285363] env[61987]: DEBUG oslo_vmware.api [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1154.285363] env[61987]: value = "task-1062426" [ 1154.285363] env[61987]: _type = "Task" [ 1154.285363] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.295970] env[61987]: DEBUG oslo_vmware.api [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062426, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.318778] env[61987]: DEBUG nova.network.neutron [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1154.452780] env[61987]: DEBUG nova.network.neutron [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Updating instance_info_cache with network_info: [{"id": "e48cc892-3cba-40a4-a803-6eab279a5d72", "address": "fa:16:3e:fb:e6:b9", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape48cc892-3c", "ovs_interfaceid": "e48cc892-3cba-40a4-a803-6eab279a5d72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.795382] env[61987]: DEBUG oslo_vmware.api [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062426, 'name': PowerOffVM_Task, 'duration_secs': 0.195476} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.797055] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1154.797055] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1154.797055] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-51a76570-3310-4c93-bbd4-4e526b13789d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.864738] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1154.865089] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1154.865365] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Deleting the datastore file [datastore2] aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1154.865814] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3f7a3fe3-8f82-4c3f-b417-3785d03afc7e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.874753] env[61987]: DEBUG oslo_vmware.api [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for the task: (returnval){ [ 1154.874753] env[61987]: value = "task-1062428" [ 1154.874753] env[61987]: _type = "Task" [ 1154.874753] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.883492] env[61987]: DEBUG oslo_vmware.api [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062428, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.956433] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Releasing lock "refresh_cache-430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1154.956709] env[61987]: DEBUG nova.compute.manager [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Instance network_info: |[{"id": "e48cc892-3cba-40a4-a803-6eab279a5d72", "address": "fa:16:3e:fb:e6:b9", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape48cc892-3c", "ovs_interfaceid": "e48cc892-3cba-40a4-a803-6eab279a5d72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1154.957219] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fb:e6:b9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'be8bd197-4b2b-46e7-88ea-2554b0438584', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e48cc892-3cba-40a4-a803-6eab279a5d72', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1154.964831] env[61987]: DEBUG oslo.service.loopingcall [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1154.965050] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1154.965277] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-03f5a048-be79-4b59-a35b-73b5a92e1971 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.985829] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1154.985829] env[61987]: value = "task-1062429" [ 1154.985829] env[61987]: _type = "Task" [ 1154.985829] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.994179] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062429, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.196281] env[61987]: DEBUG nova.compute.manager [req-30375bd6-aeed-4bc2-aed6-7965c0dd95f3 req-e27f2132-cf6a-49ae-ba3e-5620c41e37fa service nova] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Received event network-changed-e48cc892-3cba-40a4-a803-6eab279a5d72 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1155.196484] env[61987]: DEBUG nova.compute.manager [req-30375bd6-aeed-4bc2-aed6-7965c0dd95f3 req-e27f2132-cf6a-49ae-ba3e-5620c41e37fa service nova] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Refreshing instance network info cache due to event network-changed-e48cc892-3cba-40a4-a803-6eab279a5d72. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1155.196852] env[61987]: DEBUG oslo_concurrency.lockutils [req-30375bd6-aeed-4bc2-aed6-7965c0dd95f3 req-e27f2132-cf6a-49ae-ba3e-5620c41e37fa service nova] Acquiring lock "refresh_cache-430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1155.196852] env[61987]: DEBUG oslo_concurrency.lockutils [req-30375bd6-aeed-4bc2-aed6-7965c0dd95f3 req-e27f2132-cf6a-49ae-ba3e-5620c41e37fa service nova] Acquired lock "refresh_cache-430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.197028] env[61987]: DEBUG nova.network.neutron [req-30375bd6-aeed-4bc2-aed6-7965c0dd95f3 req-e27f2132-cf6a-49ae-ba3e-5620c41e37fa service nova] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Refreshing network info cache for port e48cc892-3cba-40a4-a803-6eab279a5d72 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1155.233032] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.668s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.385561] env[61987]: DEBUG oslo_vmware.api [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Task: {'id': task-1062428, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144849} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.385770] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1155.385877] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1155.386069] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1155.386245] env[61987]: INFO nova.compute.manager [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1155.386486] env[61987]: DEBUG oslo.service.loopingcall [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1155.386685] env[61987]: DEBUG nova.compute.manager [-] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1155.386785] env[61987]: DEBUG nova.network.neutron [-] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1155.496525] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062429, 'name': CreateVM_Task, 'duration_secs': 0.325464} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.496728] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1155.497400] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1155.497573] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1155.498023] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1155.498308] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ffda948-3fd2-4b14-9146-c8c4f5bf3e7c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.503515] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1155.503515] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]527f61e6-b5d0-f5e6-0125-4fd76f9c4813" [ 1155.503515] env[61987]: _type = "Task" [ 1155.503515] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.511824] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]527f61e6-b5d0-f5e6-0125-4fd76f9c4813, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.906977] env[61987]: DEBUG nova.network.neutron [req-30375bd6-aeed-4bc2-aed6-7965c0dd95f3 req-e27f2132-cf6a-49ae-ba3e-5620c41e37fa service nova] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Updated VIF entry in instance network info cache for port e48cc892-3cba-40a4-a803-6eab279a5d72. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1155.907395] env[61987]: DEBUG nova.network.neutron [req-30375bd6-aeed-4bc2-aed6-7965c0dd95f3 req-e27f2132-cf6a-49ae-ba3e-5620c41e37fa service nova] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Updating instance_info_cache with network_info: [{"id": "e48cc892-3cba-40a4-a803-6eab279a5d72", "address": "fa:16:3e:fb:e6:b9", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape48cc892-3c", "ovs_interfaceid": "e48cc892-3cba-40a4-a803-6eab279a5d72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.015192] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]527f61e6-b5d0-f5e6-0125-4fd76f9c4813, 'name': SearchDatastore_Task, 'duration_secs': 0.012691} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.015535] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1156.015775] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1156.016027] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1156.016189] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.016373] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1156.016643] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b207aab1-1593-48a7-b57a-cb47a84e2ed0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.025202] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1156.025399] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1156.026187] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4148d04-f5ac-4cf0-bd66-0a905b7c9e9b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.031480] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1156.031480] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52aab516-3839-cef6-db20-d1d7f342920b" [ 1156.031480] env[61987]: _type = "Task" [ 1156.031480] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.039649] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52aab516-3839-cef6-db20-d1d7f342920b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.174813] env[61987]: DEBUG nova.network.neutron [-] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.399865] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.400188] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.400409] env[61987]: DEBUG nova.objects.instance [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lazy-loading 'pci_requests' on Instance uuid f030044b-bc6b-478c-8f63-d519aa76203c {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1156.409875] env[61987]: DEBUG oslo_concurrency.lockutils [req-30375bd6-aeed-4bc2-aed6-7965c0dd95f3 req-e27f2132-cf6a-49ae-ba3e-5620c41e37fa service nova] Releasing lock "refresh_cache-430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1156.542944] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52aab516-3839-cef6-db20-d1d7f342920b, 'name': SearchDatastore_Task, 'duration_secs': 0.008489} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.543780] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acb685af-e544-4d97-94fb-ee98cace8b98 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.549716] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1156.549716] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5262f0c0-ac4c-2ff1-785d-ab2b12c2b4df" [ 1156.549716] env[61987]: _type = "Task" [ 1156.549716] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.558166] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5262f0c0-ac4c-2ff1-785d-ab2b12c2b4df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.677510] env[61987]: INFO nova.compute.manager [-] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Took 1.29 seconds to deallocate network for instance. [ 1156.767296] env[61987]: INFO nova.compute.manager [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Swapping old allocation on dict_keys(['8db744e5-1e14-4e98-9818-695ea4b7dfd0']) held by migration bf6d9417-391f-4b9b-8181-6d334abb9f70 for instance [ 1156.789176] env[61987]: DEBUG nova.scheduler.client.report [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Overwriting current allocation {'allocations': {'8db744e5-1e14-4e98-9818-695ea4b7dfd0': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 143}}, 'project_id': '764082416d314c3f92eb83f576aef222', 'user_id': '2941054f63934469bf4daa6b5a912e14', 'consumer_generation': 1} on consumer 5912b687-2150-4133-a3a1-44ce38b3c3cb {{(pid=61987) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 1156.878095] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "refresh_cache-5912b687-2150-4133-a3a1-44ce38b3c3cb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1156.878308] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "refresh_cache-5912b687-2150-4133-a3a1-44ce38b3c3cb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.878494] env[61987]: DEBUG nova.network.neutron [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1156.903628] env[61987]: DEBUG nova.objects.instance [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lazy-loading 'numa_topology' on Instance uuid f030044b-bc6b-478c-8f63-d519aa76203c {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1157.060612] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5262f0c0-ac4c-2ff1-785d-ab2b12c2b4df, 'name': SearchDatastore_Task, 'duration_secs': 0.01064} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.061022] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1157.061022] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3/430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1157.061250] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e033830e-31ee-4cc8-8c1f-8843d2ea16c4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.069373] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1157.069373] env[61987]: value = "task-1062430" [ 1157.069373] env[61987]: _type = "Task" [ 1157.069373] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.077862] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062430, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.184899] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1157.222478] env[61987]: DEBUG nova.compute.manager [req-90d31066-1616-42df-abe4-9166f8f3c9be req-9bf614c3-463b-4976-8611-66ac5142a7a6 service nova] [instance: aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8] Received event network-vif-deleted-70bf0473-599e-4d1e-be30-efec5a322e37 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1157.406060] env[61987]: INFO nova.compute.claims [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1157.582278] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062430, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.462853} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.582586] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3/430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1157.582809] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1157.583083] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dde72c2b-ef92-4a7f-8898-efa32a48cecd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.589804] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1157.589804] env[61987]: value = "task-1062431" [ 1157.589804] env[61987]: _type = "Task" [ 1157.589804] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.597838] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062431, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.610140] env[61987]: DEBUG nova.network.neutron [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Updating instance_info_cache with network_info: [{"id": "a8217ebd-b52f-4664-a191-dc7f3513a366", "address": "fa:16:3e:49:1b:98", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8217ebd-b5", "ovs_interfaceid": "a8217ebd-b52f-4664-a191-dc7f3513a366", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1158.100265] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062431, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06201} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.100634] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1158.101373] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2fff2bd-ed62-4061-8107-49e88a58e174 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.115013] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "refresh_cache-5912b687-2150-4133-a3a1-44ce38b3c3cb" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1158.115444] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1158.124277] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Reconfiguring VM instance instance-00000074 to attach disk [datastore1] 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3/430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1158.124519] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d4d43b5b-3919-48dc-9a1b-988cbc01785e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.126207] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a83f39f-8fa2-4370-ac04-4c97b387980f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.144984] env[61987]: DEBUG oslo_vmware.api [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1158.144984] env[61987]: value = "task-1062432" [ 1158.144984] env[61987]: _type = "Task" [ 1158.144984] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.146206] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1158.146206] env[61987]: value = "task-1062433" [ 1158.146206] env[61987]: _type = "Task" [ 1158.146206] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.156737] env[61987]: DEBUG oslo_vmware.api [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062432, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.159832] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062433, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.495629] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02fbe326-2a9b-4ddd-80c8-c13c430018e6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.502840] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d8f0ecc-d966-47a2-98be-8401d0a22a5c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.533509] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eabcb0f-a7df-4b52-8b5f-5d946cb7f047 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.541435] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcebaa37-1351-474b-bf9b-86243eb13a1c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.554787] env[61987]: DEBUG nova.compute.provider_tree [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1158.658952] env[61987]: DEBUG oslo_vmware.api [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062432, 'name': PowerOffVM_Task, 'duration_secs': 0.19077} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.661348] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1158.661986] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1158.662226] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1158.662392] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1158.662584] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1158.662739] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1158.662895] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1158.663114] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1158.663313] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1158.663510] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1158.663721] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1158.663939] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1158.668948] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062433, 'name': ReconfigVM_Task, 'duration_secs': 0.283406} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.669191] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e2edaf3-dd7a-4294-9f92-222a61b8b402 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.679033] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Reconfigured VM instance instance-00000074 to attach disk [datastore1] 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3/430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1158.679622] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-306ce8eb-cc86-4f8e-89f1-c75b78b1bfd6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.687324] env[61987]: DEBUG oslo_vmware.api [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1158.687324] env[61987]: value = "task-1062435" [ 1158.687324] env[61987]: _type = "Task" [ 1158.687324] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.688590] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1158.688590] env[61987]: value = "task-1062434" [ 1158.688590] env[61987]: _type = "Task" [ 1158.688590] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.699686] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062434, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.702620] env[61987]: DEBUG oslo_vmware.api [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062435, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.058425] env[61987]: DEBUG nova.scheduler.client.report [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1159.204265] env[61987]: DEBUG oslo_vmware.api [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062435, 'name': ReconfigVM_Task, 'duration_secs': 0.36904} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.204863] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062434, 'name': Rename_Task, 'duration_secs': 0.15328} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.205237] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1159.206026] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54177325-521c-4a04-9847-79953ec45e2f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.208449] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ce568348-2c5f-4dee-8d76-e8cd1b3229b1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.226726] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1159.227030] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1159.227154] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1159.227351] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1159.227507] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1159.227662] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1159.227904] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1159.228090] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1159.228270] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1159.228441] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1159.228628] env[61987]: DEBUG nova.virt.hardware [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1159.230625] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bce63aca-6481-4712-9b07-094c61a41d9a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.232995] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1159.232995] env[61987]: value = "task-1062436" [ 1159.232995] env[61987]: _type = "Task" [ 1159.232995] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.248110] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062436, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.249244] env[61987]: DEBUG oslo_vmware.api [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1159.249244] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5277fae9-541c-a15d-cf25-c15fcf8db6a6" [ 1159.249244] env[61987]: _type = "Task" [ 1159.249244] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.258464] env[61987]: DEBUG oslo_vmware.api [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5277fae9-541c-a15d-cf25-c15fcf8db6a6, 'name': SearchDatastore_Task, 'duration_secs': 0.01176} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.263684] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Reconfiguring VM instance instance-00000072 to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1159.263971] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23590f0f-9b8c-47fe-8c9e-9a994cc9f3b1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.283906] env[61987]: DEBUG oslo_vmware.api [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1159.283906] env[61987]: value = "task-1062437" [ 1159.283906] env[61987]: _type = "Task" [ 1159.283906] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.293451] env[61987]: DEBUG oslo_vmware.api [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062437, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.563334] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.163s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.565666] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.381s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1159.565942] env[61987]: DEBUG nova.objects.instance [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lazy-loading 'resources' on Instance uuid aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1159.594213] env[61987]: INFO nova.network.neutron [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Updating port 2d38bac7-bebb-462a-aac9-bbff9d711577 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1159.744207] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062436, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.794622] env[61987]: DEBUG oslo_vmware.api [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062437, 'name': ReconfigVM_Task, 'duration_secs': 0.216047} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.795024] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Reconfigured VM instance instance-00000072 to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1159.796021] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9cceb3a-9426-4173-a2ab-acb95f1f973d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.819908] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] 5912b687-2150-4133-a3a1-44ce38b3c3cb/5912b687-2150-4133-a3a1-44ce38b3c3cb.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1159.819908] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-98705faa-ac1c-4f54-a0fd-d5c9308fd070 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.842397] env[61987]: DEBUG oslo_vmware.api [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1159.842397] env[61987]: value = "task-1062438" [ 1159.842397] env[61987]: _type = "Task" [ 1159.842397] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.849847] env[61987]: DEBUG oslo_vmware.api [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062438, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.145738] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe2047b9-3eaa-4b8e-8d9c-5427b50c662d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.153548] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3098a216-59eb-4875-96f5-1c01a6d1ec17 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.184858] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-590a01a1-38d7-4490-a802-5d65ebc4f343 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.192510] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9d2247f-09d6-4aa1-a380-4bd6967dee7e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.206080] env[61987]: DEBUG nova.compute.provider_tree [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1160.243132] env[61987]: DEBUG oslo_vmware.api [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062436, 'name': PowerOnVM_Task, 'duration_secs': 0.552036} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.243381] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1160.243585] env[61987]: INFO nova.compute.manager [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Took 7.37 seconds to spawn the instance on the hypervisor. [ 1160.243772] env[61987]: DEBUG nova.compute.manager [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1160.244515] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f6bbe2-93f2-4677-9374-2c474bd24a0d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.350560] env[61987]: DEBUG oslo_vmware.api [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062438, 'name': ReconfigVM_Task, 'duration_secs': 0.273893} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.350791] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Reconfigured VM instance instance-00000072 to attach disk [datastore2] 5912b687-2150-4133-a3a1-44ce38b3c3cb/5912b687-2150-4133-a3a1-44ce38b3c3cb.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1160.351634] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2878a1cb-b977-4d53-ba05-4b66c03834e6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.370358] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5256712b-53b8-4414-95cc-3e44bf300eb4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.387487] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03dcb8dd-d337-4071-9d36-f2a5b108de98 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.404933] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acbaa4ec-aa01-44ad-9296-0435b63edf3b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.411454] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1160.411680] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-735fbcff-318c-4495-87d1-dbae287e25ae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.417828] env[61987]: DEBUG oslo_vmware.api [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1160.417828] env[61987]: value = "task-1062439" [ 1160.417828] env[61987]: _type = "Task" [ 1160.417828] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.425294] env[61987]: DEBUG oslo_vmware.api [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062439, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.709530] env[61987]: DEBUG nova.scheduler.client.report [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1160.762374] env[61987]: INFO nova.compute.manager [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Took 13.74 seconds to build instance. [ 1160.928913] env[61987]: DEBUG oslo_vmware.api [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062439, 'name': PowerOnVM_Task, 'duration_secs': 0.376585} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.929229] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1161.214478] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.649s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.235253] env[61987]: INFO nova.scheduler.client.report [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Deleted allocations for instance aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8 [ 1161.264629] env[61987]: DEBUG oslo_concurrency.lockutils [None req-51e67f21-472e-430e-8c03-a2333f3ccad3 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.253s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.318226] env[61987]: DEBUG nova.compute.manager [req-077d00ef-61e1-45e7-981a-f51532c8e949 req-0f8fbce5-66bb-447e-8881-7390b9221ff7 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Received event network-vif-plugged-2d38bac7-bebb-462a-aac9-bbff9d711577 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1161.318452] env[61987]: DEBUG oslo_concurrency.lockutils [req-077d00ef-61e1-45e7-981a-f51532c8e949 req-0f8fbce5-66bb-447e-8881-7390b9221ff7 service nova] Acquiring lock "f030044b-bc6b-478c-8f63-d519aa76203c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.318671] env[61987]: DEBUG oslo_concurrency.lockutils [req-077d00ef-61e1-45e7-981a-f51532c8e949 req-0f8fbce5-66bb-447e-8881-7390b9221ff7 service nova] Lock "f030044b-bc6b-478c-8f63-d519aa76203c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.318849] env[61987]: DEBUG oslo_concurrency.lockutils [req-077d00ef-61e1-45e7-981a-f51532c8e949 req-0f8fbce5-66bb-447e-8881-7390b9221ff7 service nova] Lock "f030044b-bc6b-478c-8f63-d519aa76203c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.319109] env[61987]: DEBUG nova.compute.manager [req-077d00ef-61e1-45e7-981a-f51532c8e949 req-0f8fbce5-66bb-447e-8881-7390b9221ff7 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] No waiting events found dispatching network-vif-plugged-2d38bac7-bebb-462a-aac9-bbff9d711577 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1161.319339] env[61987]: WARNING nova.compute.manager [req-077d00ef-61e1-45e7-981a-f51532c8e949 req-0f8fbce5-66bb-447e-8881-7390b9221ff7 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Received unexpected event network-vif-plugged-2d38bac7-bebb-462a-aac9-bbff9d711577 for instance with vm_state shelved_offloaded and task_state spawning. [ 1161.491980] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "refresh_cache-f030044b-bc6b-478c-8f63-d519aa76203c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1161.491980] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquired lock "refresh_cache-f030044b-bc6b-478c-8f63-d519aa76203c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1161.491980] env[61987]: DEBUG nova.network.neutron [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1161.743089] env[61987]: DEBUG oslo_concurrency.lockutils [None req-dc2ec1ea-b0e9-4b52-857f-f74286536591 tempest-ServerRescueNegativeTestJSON-1516097577 tempest-ServerRescueNegativeTestJSON-1516097577-project-member] Lock "aa15f0e3-a2ec-4f4f-bfcd-e6255f1705f8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.981s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1161.939880] env[61987]: INFO nova.compute.manager [None req-e9bed707-550e-4c8a-a519-c8a3fd8698c7 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Updating instance to original state: 'active' [ 1162.448301] env[61987]: DEBUG nova.network.neutron [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Updating instance_info_cache with network_info: [{"id": "2d38bac7-bebb-462a-aac9-bbff9d711577", "address": "fa:16:3e:f8:d4:2d", "network": {"id": "eac55241-aad5-4ddb-a01e-625c9298a8af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1411727225-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90faa728219a432fa241b635071729e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f996252-e329-42bd-a897-446dfe2b81cd", "external-id": "nsx-vlan-transportzone-535", "segmentation_id": 535, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d38bac7-be", "ovs_interfaceid": "2d38bac7-bebb-462a-aac9-bbff9d711577", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1162.956090] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Releasing lock "refresh_cache-f030044b-bc6b-478c-8f63-d519aa76203c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1162.986024] env[61987]: DEBUG nova.virt.hardware [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='292435b18a03628f8bc93bbdca970b06',container_format='bare',created_at=2024-09-30T23:17:30Z,direct_url=,disk_format='vmdk',id=9428b701-5c62-47e3-b19a-c8b536d18204,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1709286269-shelved',owner='90faa728219a432fa241b635071729e5',properties=ImageMetaProps,protected=,size=31667712,status='active',tags=,updated_at=2024-09-30T23:17:45Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1162.986219] env[61987]: DEBUG nova.virt.hardware [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1162.986323] env[61987]: DEBUG nova.virt.hardware [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1162.986548] env[61987]: DEBUG nova.virt.hardware [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1162.986728] env[61987]: DEBUG nova.virt.hardware [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1162.986894] env[61987]: DEBUG nova.virt.hardware [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1162.987158] env[61987]: DEBUG nova.virt.hardware [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1162.987347] env[61987]: DEBUG nova.virt.hardware [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1162.987539] env[61987]: DEBUG nova.virt.hardware [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1162.987737] env[61987]: DEBUG nova.virt.hardware [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1162.988010] env[61987]: DEBUG nova.virt.hardware [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1162.989578] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf1b505-5f2b-49ad-8ba1-660d87126505 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.005141] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63ba95dd-20cd-4eba-befb-a583503bd666 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.018079] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f8:d4:2d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1f996252-e329-42bd-a897-446dfe2b81cd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2d38bac7-bebb-462a-aac9-bbff9d711577', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1163.026229] env[61987]: DEBUG oslo.service.loopingcall [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1163.026722] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1163.027084] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca2c9408-0c63-41aa-b937-f7dd04473367 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.051023] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1163.051023] env[61987]: value = "task-1062441" [ 1163.051023] env[61987]: _type = "Task" [ 1163.051023] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.058205] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062441, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.226773] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "5912b687-2150-4133-a3a1-44ce38b3c3cb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.226773] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "5912b687-2150-4133-a3a1-44ce38b3c3cb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.226979] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "5912b687-2150-4133-a3a1-44ce38b3c3cb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.227102] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "5912b687-2150-4133-a3a1-44ce38b3c3cb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.227290] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "5912b687-2150-4133-a3a1-44ce38b3c3cb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.229795] env[61987]: INFO nova.compute.manager [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Terminating instance [ 1163.350214] env[61987]: DEBUG nova.compute.manager [req-9f79f6d2-451b-42cb-a0dd-f7a549ff6646 req-5a30d5ef-57f6-4160-b962-7f5c6b9fa805 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Received event network-changed-2d38bac7-bebb-462a-aac9-bbff9d711577 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1163.350447] env[61987]: DEBUG nova.compute.manager [req-9f79f6d2-451b-42cb-a0dd-f7a549ff6646 req-5a30d5ef-57f6-4160-b962-7f5c6b9fa805 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Refreshing instance network info cache due to event network-changed-2d38bac7-bebb-462a-aac9-bbff9d711577. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1163.350665] env[61987]: DEBUG oslo_concurrency.lockutils [req-9f79f6d2-451b-42cb-a0dd-f7a549ff6646 req-5a30d5ef-57f6-4160-b962-7f5c6b9fa805 service nova] Acquiring lock "refresh_cache-f030044b-bc6b-478c-8f63-d519aa76203c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1163.350816] env[61987]: DEBUG oslo_concurrency.lockutils [req-9f79f6d2-451b-42cb-a0dd-f7a549ff6646 req-5a30d5ef-57f6-4160-b962-7f5c6b9fa805 service nova] Acquired lock "refresh_cache-f030044b-bc6b-478c-8f63-d519aa76203c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1163.350988] env[61987]: DEBUG nova.network.neutron [req-9f79f6d2-451b-42cb-a0dd-f7a549ff6646 req-5a30d5ef-57f6-4160-b962-7f5c6b9fa805 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Refreshing network info cache for port 2d38bac7-bebb-462a-aac9-bbff9d711577 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1163.561028] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062441, 'name': CreateVM_Task} progress is 25%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.734186] env[61987]: DEBUG nova.compute.manager [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1163.734452] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1163.735422] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d5e91a8-dcf5-4564-bf95-e5c12a8a6050 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.744286] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1163.744553] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bf011133-d928-4900-8c62-d80a7b58a8b6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.751197] env[61987]: DEBUG oslo_vmware.api [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1163.751197] env[61987]: value = "task-1062442" [ 1163.751197] env[61987]: _type = "Task" [ 1163.751197] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.760642] env[61987]: DEBUG oslo_vmware.api [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062442, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.063035] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062441, 'name': CreateVM_Task, 'duration_secs': 0.662927} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.063035] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1164.063035] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9428b701-5c62-47e3-b19a-c8b536d18204" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1164.063035] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9428b701-5c62-47e3-b19a-c8b536d18204" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.063035] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9428b701-5c62-47e3-b19a-c8b536d18204" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1164.063035] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-658df650-3850-4a71-86ae-a1adcf9cca33 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.067267] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1164.067267] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5236e2ce-0ecb-df7b-1905-6302a54fc861" [ 1164.067267] env[61987]: _type = "Task" [ 1164.067267] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.077732] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5236e2ce-0ecb-df7b-1905-6302a54fc861, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.216385] env[61987]: DEBUG nova.network.neutron [req-9f79f6d2-451b-42cb-a0dd-f7a549ff6646 req-5a30d5ef-57f6-4160-b962-7f5c6b9fa805 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Updated VIF entry in instance network info cache for port 2d38bac7-bebb-462a-aac9-bbff9d711577. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1164.216773] env[61987]: DEBUG nova.network.neutron [req-9f79f6d2-451b-42cb-a0dd-f7a549ff6646 req-5a30d5ef-57f6-4160-b962-7f5c6b9fa805 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Updating instance_info_cache with network_info: [{"id": "2d38bac7-bebb-462a-aac9-bbff9d711577", "address": "fa:16:3e:f8:d4:2d", "network": {"id": "eac55241-aad5-4ddb-a01e-625c9298a8af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1411727225-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90faa728219a432fa241b635071729e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f996252-e329-42bd-a897-446dfe2b81cd", "external-id": "nsx-vlan-transportzone-535", "segmentation_id": 535, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d38bac7-be", "ovs_interfaceid": "2d38bac7-bebb-462a-aac9-bbff9d711577", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1164.263258] env[61987]: DEBUG oslo_vmware.api [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062442, 'name': PowerOffVM_Task, 'duration_secs': 0.228591} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.263535] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1164.263708] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1164.263961] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0125c804-8a15-486e-87aa-082faaf5dbfd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.332753] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1164.333035] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1164.333239] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Deleting the datastore file [datastore2] 5912b687-2150-4133-a3a1-44ce38b3c3cb {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1164.333516] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-362fbd72-5a62-46d8-80f8-67c41f47acdd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.340690] env[61987]: DEBUG oslo_vmware.api [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1164.340690] env[61987]: value = "task-1062444" [ 1164.340690] env[61987]: _type = "Task" [ 1164.340690] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.349302] env[61987]: DEBUG oslo_vmware.api [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062444, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.580115] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9428b701-5c62-47e3-b19a-c8b536d18204" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1164.580428] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Processing image 9428b701-5c62-47e3-b19a-c8b536d18204 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1164.580624] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9428b701-5c62-47e3-b19a-c8b536d18204/9428b701-5c62-47e3-b19a-c8b536d18204.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1164.580780] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9428b701-5c62-47e3-b19a-c8b536d18204/9428b701-5c62-47e3-b19a-c8b536d18204.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.580961] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1164.581230] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8e8374cc-0e6a-4cf3-8dfe-7b49aac7fd6b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.589623] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1164.589817] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1164.590636] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db46fbec-2b1f-4df7-be48-791d5254bf34 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.595827] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1164.595827] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52d1779c-f790-788a-a27e-26cbc76ecdc9" [ 1164.595827] env[61987]: _type = "Task" [ 1164.595827] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.603526] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52d1779c-f790-788a-a27e-26cbc76ecdc9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.720539] env[61987]: DEBUG oslo_concurrency.lockutils [req-9f79f6d2-451b-42cb-a0dd-f7a549ff6646 req-5a30d5ef-57f6-4160-b962-7f5c6b9fa805 service nova] Releasing lock "refresh_cache-f030044b-bc6b-478c-8f63-d519aa76203c" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1164.720837] env[61987]: DEBUG nova.compute.manager [req-9f79f6d2-451b-42cb-a0dd-f7a549ff6646 req-5a30d5ef-57f6-4160-b962-7f5c6b9fa805 service nova] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Received event network-changed-e48cc892-3cba-40a4-a803-6eab279a5d72 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1164.721031] env[61987]: DEBUG nova.compute.manager [req-9f79f6d2-451b-42cb-a0dd-f7a549ff6646 req-5a30d5ef-57f6-4160-b962-7f5c6b9fa805 service nova] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Refreshing instance network info cache due to event network-changed-e48cc892-3cba-40a4-a803-6eab279a5d72. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1164.721268] env[61987]: DEBUG oslo_concurrency.lockutils [req-9f79f6d2-451b-42cb-a0dd-f7a549ff6646 req-5a30d5ef-57f6-4160-b962-7f5c6b9fa805 service nova] Acquiring lock "refresh_cache-430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1164.721418] env[61987]: DEBUG oslo_concurrency.lockutils [req-9f79f6d2-451b-42cb-a0dd-f7a549ff6646 req-5a30d5ef-57f6-4160-b962-7f5c6b9fa805 service nova] Acquired lock "refresh_cache-430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.721590] env[61987]: DEBUG nova.network.neutron [req-9f79f6d2-451b-42cb-a0dd-f7a549ff6646 req-5a30d5ef-57f6-4160-b962-7f5c6b9fa805 service nova] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Refreshing network info cache for port e48cc892-3cba-40a4-a803-6eab279a5d72 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1164.851288] env[61987]: DEBUG oslo_vmware.api [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062444, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150218} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.851288] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1164.851549] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1164.851549] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1164.851722] env[61987]: INFO nova.compute.manager [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1164.852268] env[61987]: DEBUG oslo.service.loopingcall [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1164.852268] env[61987]: DEBUG nova.compute.manager [-] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1164.852268] env[61987]: DEBUG nova.network.neutron [-] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1165.107885] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Preparing fetch location {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1165.108427] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Fetch image to [datastore2] OSTACK_IMG_886e78a0-75c2-4c44-a6c0-e4237497b041/OSTACK_IMG_886e78a0-75c2-4c44-a6c0-e4237497b041.vmdk {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1165.108536] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Downloading stream optimized image 9428b701-5c62-47e3-b19a-c8b536d18204 to [datastore2] OSTACK_IMG_886e78a0-75c2-4c44-a6c0-e4237497b041/OSTACK_IMG_886e78a0-75c2-4c44-a6c0-e4237497b041.vmdk on the data store datastore2 as vApp {{(pid=61987) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1165.108692] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Downloading image file data 9428b701-5c62-47e3-b19a-c8b536d18204 to the ESX as VM named 'OSTACK_IMG_886e78a0-75c2-4c44-a6c0-e4237497b041' {{(pid=61987) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1165.182449] env[61987]: DEBUG oslo_vmware.rw_handles [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1165.182449] env[61987]: value = "resgroup-9" [ 1165.182449] env[61987]: _type = "ResourcePool" [ 1165.182449] env[61987]: }. {{(pid=61987) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1165.182756] env[61987]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-424a15a7-9941-42d8-b88a-3297ffa2a5df {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.213832] env[61987]: DEBUG oslo_vmware.rw_handles [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lease: (returnval){ [ 1165.213832] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]525f9227-1d5e-d5a7-1ffe-5fe59f1a4252" [ 1165.213832] env[61987]: _type = "HttpNfcLease" [ 1165.213832] env[61987]: } obtained for vApp import into resource pool (val){ [ 1165.213832] env[61987]: value = "resgroup-9" [ 1165.213832] env[61987]: _type = "ResourcePool" [ 1165.213832] env[61987]: }. {{(pid=61987) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1165.214452] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the lease: (returnval){ [ 1165.214452] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]525f9227-1d5e-d5a7-1ffe-5fe59f1a4252" [ 1165.214452] env[61987]: _type = "HttpNfcLease" [ 1165.214452] env[61987]: } to be ready. {{(pid=61987) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1165.227615] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1165.227615] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]525f9227-1d5e-d5a7-1ffe-5fe59f1a4252" [ 1165.227615] env[61987]: _type = "HttpNfcLease" [ 1165.227615] env[61987]: } is initializing. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1165.378521] env[61987]: DEBUG nova.compute.manager [req-6f579e01-f897-4d66-ab31-aa3433bc8cfd req-7e5c7ed0-4585-43e1-b1cc-3a727c89bfcb service nova] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Received event network-vif-deleted-a8217ebd-b52f-4664-a191-dc7f3513a366 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1165.378780] env[61987]: INFO nova.compute.manager [req-6f579e01-f897-4d66-ab31-aa3433bc8cfd req-7e5c7ed0-4585-43e1-b1cc-3a727c89bfcb service nova] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Neutron deleted interface a8217ebd-b52f-4664-a191-dc7f3513a366; detaching it from the instance and deleting it from the info cache [ 1165.378945] env[61987]: DEBUG nova.network.neutron [req-6f579e01-f897-4d66-ab31-aa3433bc8cfd req-7e5c7ed0-4585-43e1-b1cc-3a727c89bfcb service nova] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1165.454419] env[61987]: DEBUG nova.network.neutron [req-9f79f6d2-451b-42cb-a0dd-f7a549ff6646 req-5a30d5ef-57f6-4160-b962-7f5c6b9fa805 service nova] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Updated VIF entry in instance network info cache for port e48cc892-3cba-40a4-a803-6eab279a5d72. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1165.454808] env[61987]: DEBUG nova.network.neutron [req-9f79f6d2-451b-42cb-a0dd-f7a549ff6646 req-5a30d5ef-57f6-4160-b962-7f5c6b9fa805 service nova] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Updating instance_info_cache with network_info: [{"id": "e48cc892-3cba-40a4-a803-6eab279a5d72", "address": "fa:16:3e:fb:e6:b9", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape48cc892-3c", "ovs_interfaceid": "e48cc892-3cba-40a4-a803-6eab279a5d72", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1165.724334] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1165.724334] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]525f9227-1d5e-d5a7-1ffe-5fe59f1a4252" [ 1165.724334] env[61987]: _type = "HttpNfcLease" [ 1165.724334] env[61987]: } is initializing. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1165.812257] env[61987]: DEBUG nova.network.neutron [-] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1165.881487] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e435c9ad-59b5-40c5-b094-3b92ff33fde6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.892577] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d87ffc43-e4d7-4137-95ec-fb6fa448a1b1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.923895] env[61987]: DEBUG nova.compute.manager [req-6f579e01-f897-4d66-ab31-aa3433bc8cfd req-7e5c7ed0-4585-43e1-b1cc-3a727c89bfcb service nova] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Detach interface failed, port_id=a8217ebd-b52f-4664-a191-dc7f3513a366, reason: Instance 5912b687-2150-4133-a3a1-44ce38b3c3cb could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 1165.959272] env[61987]: DEBUG oslo_concurrency.lockutils [req-9f79f6d2-451b-42cb-a0dd-f7a549ff6646 req-5a30d5ef-57f6-4160-b962-7f5c6b9fa805 service nova] Releasing lock "refresh_cache-430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1166.223667] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1166.223667] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]525f9227-1d5e-d5a7-1ffe-5fe59f1a4252" [ 1166.223667] env[61987]: _type = "HttpNfcLease" [ 1166.223667] env[61987]: } is ready. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1166.224374] env[61987]: DEBUG oslo_vmware.rw_handles [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1166.224374] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]525f9227-1d5e-d5a7-1ffe-5fe59f1a4252" [ 1166.224374] env[61987]: _type = "HttpNfcLease" [ 1166.224374] env[61987]: }. {{(pid=61987) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1166.224753] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77e4472f-d577-44a1-b71c-553809d323bc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.232578] env[61987]: DEBUG oslo_vmware.rw_handles [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5210564f-85d4-c6b3-eef9-936521cf16db/disk-0.vmdk from lease info. {{(pid=61987) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1166.232767] env[61987]: DEBUG oslo_vmware.rw_handles [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Creating HTTP connection to write to file with size = 31667712 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5210564f-85d4-c6b3-eef9-936521cf16db/disk-0.vmdk. {{(pid=61987) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1166.300887] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-a51c6c09-299d-49ad-a8cc-9d81391e0c30 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.315276] env[61987]: INFO nova.compute.manager [-] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Took 1.46 seconds to deallocate network for instance. [ 1166.822508] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.822961] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.823280] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1166.854299] env[61987]: INFO nova.scheduler.client.report [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Deleted allocations for instance 5912b687-2150-4133-a3a1-44ce38b3c3cb [ 1167.363988] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae5752d1-fa63-442d-b88b-f689aae0dfb4 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "5912b687-2150-4133-a3a1-44ce38b3c3cb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.137s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.395056] env[61987]: DEBUG oslo_vmware.rw_handles [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Completed reading data from the image iterator. {{(pid=61987) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1167.395357] env[61987]: DEBUG oslo_vmware.rw_handles [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5210564f-85d4-c6b3-eef9-936521cf16db/disk-0.vmdk. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1167.396614] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3411692-da76-48ed-aee6-16ed783d1a69 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.405295] env[61987]: DEBUG oslo_vmware.rw_handles [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5210564f-85d4-c6b3-eef9-936521cf16db/disk-0.vmdk is in state: ready. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1167.405486] env[61987]: DEBUG oslo_vmware.rw_handles [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5210564f-85d4-c6b3-eef9-936521cf16db/disk-0.vmdk. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1167.405750] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-fd92ad7b-1527-4546-bd01-f6d302813673 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.606058] env[61987]: DEBUG oslo_vmware.rw_handles [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5210564f-85d4-c6b3-eef9-936521cf16db/disk-0.vmdk. {{(pid=61987) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1167.606281] env[61987]: INFO nova.virt.vmwareapi.images [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Downloaded image file data 9428b701-5c62-47e3-b19a-c8b536d18204 [ 1167.607156] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5d2fda4-98f8-43f9-ba61-bf2763c0f963 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.623350] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bec09975-9034-407b-9d5d-d325663717db {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.655374] env[61987]: INFO nova.virt.vmwareapi.images [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] The imported VM was unregistered [ 1167.658235] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Caching image {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1167.658516] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Creating directory with path [datastore2] devstack-image-cache_base/9428b701-5c62-47e3-b19a-c8b536d18204 {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1167.658959] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-861ee709-db40-4b30-bb66-2faf36db3600 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.671340] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Created directory with path [datastore2] devstack-image-cache_base/9428b701-5c62-47e3-b19a-c8b536d18204 {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1167.671562] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_886e78a0-75c2-4c44-a6c0-e4237497b041/OSTACK_IMG_886e78a0-75c2-4c44-a6c0-e4237497b041.vmdk to [datastore2] devstack-image-cache_base/9428b701-5c62-47e3-b19a-c8b536d18204/9428b701-5c62-47e3-b19a-c8b536d18204.vmdk. {{(pid=61987) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1167.671859] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-a8768a74-5f9c-48fa-b0ff-b056b9264deb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.679574] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1167.679574] env[61987]: value = "task-1062449" [ 1167.679574] env[61987]: _type = "Task" [ 1167.679574] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.690915] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062449, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.192813] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062449, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.366117] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "0704efaf-08de-4888-a092-476bcb789eea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.366527] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "0704efaf-08de-4888-a092-476bcb789eea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.695399] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062449, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.869863] env[61987]: DEBUG nova.compute.manager [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1169.154560] env[61987]: DEBUG oslo_concurrency.lockutils [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Acquiring lock "2ba77d32-d777-448e-8fa6-04920950f0f6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.154838] env[61987]: DEBUG oslo_concurrency.lockutils [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Lock "2ba77d32-d777-448e-8fa6-04920950f0f6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.155085] env[61987]: DEBUG oslo_concurrency.lockutils [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Acquiring lock "2ba77d32-d777-448e-8fa6-04920950f0f6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.155285] env[61987]: DEBUG oslo_concurrency.lockutils [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Lock "2ba77d32-d777-448e-8fa6-04920950f0f6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.155471] env[61987]: DEBUG oslo_concurrency.lockutils [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Lock "2ba77d32-d777-448e-8fa6-04920950f0f6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.157928] env[61987]: INFO nova.compute.manager [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Terminating instance [ 1169.193133] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062449, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.393258] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.393567] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.395302] env[61987]: INFO nova.compute.claims [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1169.663069] env[61987]: DEBUG nova.compute.manager [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1169.663422] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1169.663786] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-867f2cd5-07ac-413d-b483-75893b7c8c99 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.673491] env[61987]: DEBUG oslo_vmware.api [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Waiting for the task: (returnval){ [ 1169.673491] env[61987]: value = "task-1062451" [ 1169.673491] env[61987]: _type = "Task" [ 1169.673491] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.684023] env[61987]: DEBUG oslo_vmware.api [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062451, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.695325] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062449, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.184499] env[61987]: DEBUG oslo_vmware.api [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062451, 'name': PowerOffVM_Task, 'duration_secs': 0.237346} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.187621] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1170.187839] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Volume detach. Driver type: vmdk {{(pid=61987) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1170.188080] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234441', 'volume_id': '4ab1a236-de8d-4578-8274-89bc106faa2e', 'name': 'volume-4ab1a236-de8d-4578-8274-89bc106faa2e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2ba77d32-d777-448e-8fa6-04920950f0f6', 'attached_at': '', 'detached_at': '', 'volume_id': '4ab1a236-de8d-4578-8274-89bc106faa2e', 'serial': '4ab1a236-de8d-4578-8274-89bc106faa2e'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1170.188802] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e95a68f-ffae-4cb4-995f-14f9b120b151 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.196250] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062449, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.41301} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.209258] env[61987]: INFO nova.virt.vmwareapi.ds_util [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_886e78a0-75c2-4c44-a6c0-e4237497b041/OSTACK_IMG_886e78a0-75c2-4c44-a6c0-e4237497b041.vmdk to [datastore2] devstack-image-cache_base/9428b701-5c62-47e3-b19a-c8b536d18204/9428b701-5c62-47e3-b19a-c8b536d18204.vmdk. [ 1170.209457] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Cleaning up location [datastore2] OSTACK_IMG_886e78a0-75c2-4c44-a6c0-e4237497b041 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1170.209673] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_886e78a0-75c2-4c44-a6c0-e4237497b041 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1170.209986] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e4e4b696-6ad5-4668-81ae-5929a3bd61cb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.211949] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c403707f-48a1-400a-aa11-45b02b35368e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.219985] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc4a448b-0492-4567-825f-03f92c8b5f26 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.222328] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1170.222328] env[61987]: value = "task-1062452" [ 1170.222328] env[61987]: _type = "Task" [ 1170.222328] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.239953] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c1b6458-38ff-4567-bdf4-87f11f22df33 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.244999] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062452, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.257011] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] The volume has not been displaced from its original location: [datastore1] volume-4ab1a236-de8d-4578-8274-89bc106faa2e/volume-4ab1a236-de8d-4578-8274-89bc106faa2e.vmdk. No consolidation needed. {{(pid=61987) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1170.262320] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Reconfiguring VM instance instance-00000073 to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1170.262606] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b6ea66b-f489-401c-98ef-6159bbb0cffe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.280832] env[61987]: DEBUG oslo_vmware.api [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Waiting for the task: (returnval){ [ 1170.280832] env[61987]: value = "task-1062453" [ 1170.280832] env[61987]: _type = "Task" [ 1170.280832] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.289308] env[61987]: DEBUG oslo_vmware.api [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062453, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.476576] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e5cb742-ab9c-4000-8ab6-e9868e316714 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.484327] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-163de729-773f-45d5-ba9e-ba2a375ff97b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.513853] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae3976ea-ea8c-4ee6-a0de-71af1c7e8f19 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.520963] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19150c95-8cd1-4482-aab6-40a09d3aaea4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.533473] env[61987]: DEBUG nova.compute.provider_tree [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1170.734583] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062452, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.039602} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.734859] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1170.735053] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9428b701-5c62-47e3-b19a-c8b536d18204/9428b701-5c62-47e3-b19a-c8b536d18204.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1170.735330] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9428b701-5c62-47e3-b19a-c8b536d18204/9428b701-5c62-47e3-b19a-c8b536d18204.vmdk to [datastore2] f030044b-bc6b-478c-8f63-d519aa76203c/f030044b-bc6b-478c-8f63-d519aa76203c.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1170.735600] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f8cd7fc3-4687-4a6a-be73-2fe35fd1494d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.742876] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1170.742876] env[61987]: value = "task-1062454" [ 1170.742876] env[61987]: _type = "Task" [ 1170.742876] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.753223] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062454, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.791172] env[61987]: DEBUG oslo_vmware.api [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062453, 'name': ReconfigVM_Task, 'duration_secs': 0.187063} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.791485] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Reconfigured VM instance instance-00000073 to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1170.796482] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31411510-085f-4a2d-9c30-c118db1a03fe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.813637] env[61987]: DEBUG oslo_vmware.api [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Waiting for the task: (returnval){ [ 1170.813637] env[61987]: value = "task-1062455" [ 1170.813637] env[61987]: _type = "Task" [ 1170.813637] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.825705] env[61987]: DEBUG oslo_vmware.api [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062455, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.036865] env[61987]: DEBUG nova.scheduler.client.report [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1171.253750] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062454, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.324523] env[61987]: DEBUG oslo_vmware.api [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062455, 'name': ReconfigVM_Task, 'duration_secs': 0.237405} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.324863] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234441', 'volume_id': '4ab1a236-de8d-4578-8274-89bc106faa2e', 'name': 'volume-4ab1a236-de8d-4578-8274-89bc106faa2e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2ba77d32-d777-448e-8fa6-04920950f0f6', 'attached_at': '', 'detached_at': '', 'volume_id': '4ab1a236-de8d-4578-8274-89bc106faa2e', 'serial': '4ab1a236-de8d-4578-8274-89bc106faa2e'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1171.325173] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1171.326041] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c28a32d8-e983-42b0-9fa2-38049976cf21 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.334027] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1171.334308] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3752a1eb-0d1a-4028-a628-17349cda9331 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.407020] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1171.407330] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1171.407580] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Deleting the datastore file [datastore1] 2ba77d32-d777-448e-8fa6-04920950f0f6 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1171.407830] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-300d1796-09f4-4e34-b235-7e9ec743f7eb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.415494] env[61987]: DEBUG oslo_vmware.api [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Waiting for the task: (returnval){ [ 1171.415494] env[61987]: value = "task-1062457" [ 1171.415494] env[61987]: _type = "Task" [ 1171.415494] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.424415] env[61987]: DEBUG oslo_vmware.api [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062457, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.541773] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.148s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.542515] env[61987]: DEBUG nova.compute.manager [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1171.754693] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062454, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.927168] env[61987]: DEBUG oslo_vmware.api [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Task: {'id': task-1062457, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.232814} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.927470] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1171.927680] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1171.927956] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1171.928178] env[61987]: INFO nova.compute.manager [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Took 2.26 seconds to destroy the instance on the hypervisor. [ 1171.928459] env[61987]: DEBUG oslo.service.loopingcall [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1171.928674] env[61987]: DEBUG nova.compute.manager [-] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1171.928790] env[61987]: DEBUG nova.network.neutron [-] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1172.048412] env[61987]: DEBUG nova.compute.utils [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1172.050110] env[61987]: DEBUG nova.compute.manager [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1172.050254] env[61987]: DEBUG nova.network.neutron [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1172.091244] env[61987]: DEBUG nova.policy [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2941054f63934469bf4daa6b5a912e14', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '764082416d314c3f92eb83f576aef222', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 1172.258417] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062454, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.455681] env[61987]: DEBUG nova.network.neutron [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Successfully created port: 79e1d3b3-41db-4650-95bd-11df0dc06e80 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1172.554829] env[61987]: DEBUG nova.compute.manager [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1172.564879] env[61987]: DEBUG nova.compute.manager [req-d90d7161-1961-41ae-815f-2cfcf9f267ff req-be9112de-d8ff-4966-b3f3-5576b9267c30 service nova] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Received event network-vif-deleted-2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1172.566076] env[61987]: INFO nova.compute.manager [req-d90d7161-1961-41ae-815f-2cfcf9f267ff req-be9112de-d8ff-4966-b3f3-5576b9267c30 service nova] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Neutron deleted interface 2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a; detaching it from the instance and deleting it from the info cache [ 1172.566363] env[61987]: DEBUG nova.network.neutron [req-d90d7161-1961-41ae-815f-2cfcf9f267ff req-be9112de-d8ff-4966-b3f3-5576b9267c30 service nova] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.755739] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062454, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.917817] env[61987]: DEBUG nova.network.neutron [-] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1173.069401] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-52838204-71b4-4d48-8b4c-03f999b95bb5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.080563] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fcfec4d-ee77-4419-ba7b-5088fea5f538 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.110924] env[61987]: DEBUG nova.compute.manager [req-d90d7161-1961-41ae-815f-2cfcf9f267ff req-be9112de-d8ff-4966-b3f3-5576b9267c30 service nova] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Detach interface failed, port_id=2cf009f7-22f8-4da8-b5eb-6b2ed4e6572a, reason: Instance 2ba77d32-d777-448e-8fa6-04920950f0f6 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 1173.256558] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062454, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.420978] env[61987]: INFO nova.compute.manager [-] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Took 1.49 seconds to deallocate network for instance. [ 1173.569959] env[61987]: DEBUG nova.compute.manager [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1173.596205] env[61987]: DEBUG nova.virt.hardware [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1173.596467] env[61987]: DEBUG nova.virt.hardware [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1173.596647] env[61987]: DEBUG nova.virt.hardware [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1173.596850] env[61987]: DEBUG nova.virt.hardware [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1173.597018] env[61987]: DEBUG nova.virt.hardware [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1173.597192] env[61987]: DEBUG nova.virt.hardware [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1173.597408] env[61987]: DEBUG nova.virt.hardware [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1173.597579] env[61987]: DEBUG nova.virt.hardware [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1173.597752] env[61987]: DEBUG nova.virt.hardware [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1173.597919] env[61987]: DEBUG nova.virt.hardware [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1173.598150] env[61987]: DEBUG nova.virt.hardware [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1173.599020] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a91f9198-cdf4-475d-b880-1f56c2602865 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.607434] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceea612c-5cec-458f-95e3-dd4c4ee9f7f4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.757981] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062454, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.550625} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.758314] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9428b701-5c62-47e3-b19a-c8b536d18204/9428b701-5c62-47e3-b19a-c8b536d18204.vmdk to [datastore2] f030044b-bc6b-478c-8f63-d519aa76203c/f030044b-bc6b-478c-8f63-d519aa76203c.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1173.759209] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14e8347b-dda6-4d30-92e0-216241e9611d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.784260] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] f030044b-bc6b-478c-8f63-d519aa76203c/f030044b-bc6b-478c-8f63-d519aa76203c.vmdk or device None with type streamOptimized {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1173.784589] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0cd3de1-3c59-4c4c-a244-0700c79e6ac4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.805234] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1173.805234] env[61987]: value = "task-1062460" [ 1173.805234] env[61987]: _type = "Task" [ 1173.805234] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.816597] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062460, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.900077] env[61987]: DEBUG nova.compute.manager [req-ebdd2714-05d2-4ad4-b84b-637e82db1776 req-ef59dedb-7a36-4069-84fa-ea368ac902e6 service nova] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Received event network-vif-plugged-79e1d3b3-41db-4650-95bd-11df0dc06e80 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1173.900468] env[61987]: DEBUG oslo_concurrency.lockutils [req-ebdd2714-05d2-4ad4-b84b-637e82db1776 req-ef59dedb-7a36-4069-84fa-ea368ac902e6 service nova] Acquiring lock "0704efaf-08de-4888-a092-476bcb789eea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.900522] env[61987]: DEBUG oslo_concurrency.lockutils [req-ebdd2714-05d2-4ad4-b84b-637e82db1776 req-ef59dedb-7a36-4069-84fa-ea368ac902e6 service nova] Lock "0704efaf-08de-4888-a092-476bcb789eea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.900688] env[61987]: DEBUG oslo_concurrency.lockutils [req-ebdd2714-05d2-4ad4-b84b-637e82db1776 req-ef59dedb-7a36-4069-84fa-ea368ac902e6 service nova] Lock "0704efaf-08de-4888-a092-476bcb789eea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1173.900860] env[61987]: DEBUG nova.compute.manager [req-ebdd2714-05d2-4ad4-b84b-637e82db1776 req-ef59dedb-7a36-4069-84fa-ea368ac902e6 service nova] [instance: 0704efaf-08de-4888-a092-476bcb789eea] No waiting events found dispatching network-vif-plugged-79e1d3b3-41db-4650-95bd-11df0dc06e80 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1173.901045] env[61987]: WARNING nova.compute.manager [req-ebdd2714-05d2-4ad4-b84b-637e82db1776 req-ef59dedb-7a36-4069-84fa-ea368ac902e6 service nova] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Received unexpected event network-vif-plugged-79e1d3b3-41db-4650-95bd-11df0dc06e80 for instance with vm_state building and task_state spawning. [ 1173.967324] env[61987]: INFO nova.compute.manager [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Took 0.55 seconds to detach 1 volumes for instance. [ 1173.969599] env[61987]: DEBUG nova.compute.manager [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Deleting volume: 4ab1a236-de8d-4578-8274-89bc106faa2e {{(pid=61987) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3281}} [ 1173.986575] env[61987]: DEBUG nova.network.neutron [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Successfully updated port: 79e1d3b3-41db-4650-95bd-11df0dc06e80 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1174.319157] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062460, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.489789] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "refresh_cache-0704efaf-08de-4888-a092-476bcb789eea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1174.490197] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "refresh_cache-0704efaf-08de-4888-a092-476bcb789eea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1174.490197] env[61987]: DEBUG nova.network.neutron [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1174.515235] env[61987]: DEBUG oslo_concurrency.lockutils [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1174.515525] env[61987]: DEBUG oslo_concurrency.lockutils [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1174.515751] env[61987]: DEBUG nova.objects.instance [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Lazy-loading 'resources' on Instance uuid 2ba77d32-d777-448e-8fa6-04920950f0f6 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1174.817257] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062460, 'name': ReconfigVM_Task, 'duration_secs': 0.761214} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.817566] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Reconfigured VM instance instance-00000071 to attach disk [datastore2] f030044b-bc6b-478c-8f63-d519aa76203c/f030044b-bc6b-478c-8f63-d519aa76203c.vmdk or device None with type streamOptimized {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1174.818842] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'device_type': 'disk', 'device_name': '/dev/sda', 'encryption_secret_uuid': None, 'encryption_format': None, 'encryption_options': None, 'boot_index': 0, 'guest_format': None, 'size': 0, 'encrypted': False, 'disk_bus': None, 'image_id': '7bf8969f-d179-42af-9e15-527d5449f929'}], 'ephemerals': [], 'block_device_mapping': [{'device_type': None, 'attachment_id': '8d4a2c70-5bb9-4c42-b08a-953ddb3121c4', 'mount_device': '/dev/sdb', 'delete_on_termination': False, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234450', 'volume_id': 'e6316fa6-4d06-40fb-90d8-60c4cfd7c454', 'name': 'volume-e6316fa6-4d06-40fb-90d8-60c4cfd7c454', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'f030044b-bc6b-478c-8f63-d519aa76203c', 'attached_at': '', 'detached_at': '', 'volume_id': 'e6316fa6-4d06-40fb-90d8-60c4cfd7c454', 'serial': 'e6316fa6-4d06-40fb-90d8-60c4cfd7c454'}, 'boot_index': None, 'guest_format': None, 'disk_bus': None, 'volume_type': None}], 'swap': None} {{(pid=61987) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1174.819121] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Volume attach. Driver type: vmdk {{(pid=61987) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1174.819354] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234450', 'volume_id': 'e6316fa6-4d06-40fb-90d8-60c4cfd7c454', 'name': 'volume-e6316fa6-4d06-40fb-90d8-60c4cfd7c454', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'f030044b-bc6b-478c-8f63-d519aa76203c', 'attached_at': '', 'detached_at': '', 'volume_id': 'e6316fa6-4d06-40fb-90d8-60c4cfd7c454', 'serial': 'e6316fa6-4d06-40fb-90d8-60c4cfd7c454'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1174.820168] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-895cb11c-7ec5-40e3-99bf-406ea796059c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.836280] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e02a6cb-3a6e-4794-b4c8-5fee67339ef9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.860768] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Reconfiguring VM instance instance-00000071 to attach disk [datastore1] volume-e6316fa6-4d06-40fb-90d8-60c4cfd7c454/volume-e6316fa6-4d06-40fb-90d8-60c4cfd7c454.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1174.861090] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5f2a964-6da2-4317-b642-61394f928cac {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.880203] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1174.880203] env[61987]: value = "task-1062462" [ 1174.880203] env[61987]: _type = "Task" [ 1174.880203] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.888288] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062462, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.021251] env[61987]: DEBUG nova.network.neutron [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1175.102720] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf33f9d-563c-4d97-a687-3bc7b721bfc7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.110824] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-179c39b4-6fa1-49a4-8c13-d15ffa559717 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.148886] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b0fb5ef-a9ef-4427-a3e4-2089b6461a74 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.156550] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d52a3e4d-9d3d-4074-b978-7b780f01dfd1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.170752] env[61987]: DEBUG nova.compute.provider_tree [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1175.187027] env[61987]: DEBUG nova.network.neutron [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Updating instance_info_cache with network_info: [{"id": "79e1d3b3-41db-4650-95bd-11df0dc06e80", "address": "fa:16:3e:7f:09:fd", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79e1d3b3-41", "ovs_interfaceid": "79e1d3b3-41db-4650-95bd-11df0dc06e80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1175.392027] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062462, 'name': ReconfigVM_Task, 'duration_secs': 0.303465} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.392159] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Reconfigured VM instance instance-00000071 to attach disk [datastore1] volume-e6316fa6-4d06-40fb-90d8-60c4cfd7c454/volume-e6316fa6-4d06-40fb-90d8-60c4cfd7c454.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1175.397146] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97fe4547-19b0-4bcc-869f-f3c54fa6e2cd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.413677] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1175.413677] env[61987]: value = "task-1062463" [ 1175.413677] env[61987]: _type = "Task" [ 1175.413677] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.422418] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062463, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.673757] env[61987]: DEBUG nova.scheduler.client.report [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1175.689279] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "refresh_cache-0704efaf-08de-4888-a092-476bcb789eea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1175.689579] env[61987]: DEBUG nova.compute.manager [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Instance network_info: |[{"id": "79e1d3b3-41db-4650-95bd-11df0dc06e80", "address": "fa:16:3e:7f:09:fd", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79e1d3b3-41", "ovs_interfaceid": "79e1d3b3-41db-4650-95bd-11df0dc06e80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1175.690025] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:09:fd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2a111da1-447a-4722-9662-b0d6a2886a65', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '79e1d3b3-41db-4650-95bd-11df0dc06e80', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1175.697403] env[61987]: DEBUG oslo.service.loopingcall [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1175.697923] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1175.698212] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d00ce891-4a0e-47cc-ab77-58d5ec4fd990 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.719051] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1175.719051] env[61987]: value = "task-1062464" [ 1175.719051] env[61987]: _type = "Task" [ 1175.719051] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.726978] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062464, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.925856] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062463, 'name': ReconfigVM_Task, 'duration_secs': 0.150509} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.927318] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234450', 'volume_id': 'e6316fa6-4d06-40fb-90d8-60c4cfd7c454', 'name': 'volume-e6316fa6-4d06-40fb-90d8-60c4cfd7c454', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'f030044b-bc6b-478c-8f63-d519aa76203c', 'attached_at': '', 'detached_at': '', 'volume_id': 'e6316fa6-4d06-40fb-90d8-60c4cfd7c454', 'serial': 'e6316fa6-4d06-40fb-90d8-60c4cfd7c454'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1175.928796] env[61987]: DEBUG nova.compute.manager [req-986d48a6-aaa5-4033-bd3f-09a01ce7d8f2 req-869ec86a-d4ad-46ec-8723-e78ca91246b7 service nova] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Received event network-changed-79e1d3b3-41db-4650-95bd-11df0dc06e80 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1175.929019] env[61987]: DEBUG nova.compute.manager [req-986d48a6-aaa5-4033-bd3f-09a01ce7d8f2 req-869ec86a-d4ad-46ec-8723-e78ca91246b7 service nova] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Refreshing instance network info cache due to event network-changed-79e1d3b3-41db-4650-95bd-11df0dc06e80. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1175.929321] env[61987]: DEBUG oslo_concurrency.lockutils [req-986d48a6-aaa5-4033-bd3f-09a01ce7d8f2 req-869ec86a-d4ad-46ec-8723-e78ca91246b7 service nova] Acquiring lock "refresh_cache-0704efaf-08de-4888-a092-476bcb789eea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1175.929547] env[61987]: DEBUG oslo_concurrency.lockutils [req-986d48a6-aaa5-4033-bd3f-09a01ce7d8f2 req-869ec86a-d4ad-46ec-8723-e78ca91246b7 service nova] Acquired lock "refresh_cache-0704efaf-08de-4888-a092-476bcb789eea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.929783] env[61987]: DEBUG nova.network.neutron [req-986d48a6-aaa5-4033-bd3f-09a01ce7d8f2 req-869ec86a-d4ad-46ec-8723-e78ca91246b7 service nova] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Refreshing network info cache for port 79e1d3b3-41db-4650-95bd-11df0dc06e80 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1175.931037] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fdffef98-49bb-4728-848d-07d06da6e849 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.939184] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1175.939184] env[61987]: value = "task-1062466" [ 1175.939184] env[61987]: _type = "Task" [ 1175.939184] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.952056] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062466, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.178780] env[61987]: DEBUG oslo_concurrency.lockutils [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.663s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.196543] env[61987]: INFO nova.scheduler.client.report [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Deleted allocations for instance 2ba77d32-d777-448e-8fa6-04920950f0f6 [ 1176.231839] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062464, 'name': CreateVM_Task, 'duration_secs': 0.332985} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.231839] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1176.232045] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1176.232303] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.232638] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1176.232897] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2a627230-38b3-4947-a6df-2b52fad899af {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.241111] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1176.241111] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]523ee3f9-d8aa-cfa7-26dc-9d39b71beda7" [ 1176.241111] env[61987]: _type = "Task" [ 1176.241111] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.249815] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523ee3f9-d8aa-cfa7-26dc-9d39b71beda7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.450186] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062466, 'name': Rename_Task, 'duration_secs': 0.188696} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.450517] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1176.450825] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf1b6579-2740-4962-b620-66880357ba93 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.458595] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1176.458595] env[61987]: value = "task-1062467" [ 1176.458595] env[61987]: _type = "Task" [ 1176.458595] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.468286] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062467, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.686231] env[61987]: DEBUG nova.network.neutron [req-986d48a6-aaa5-4033-bd3f-09a01ce7d8f2 req-869ec86a-d4ad-46ec-8723-e78ca91246b7 service nova] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Updated VIF entry in instance network info cache for port 79e1d3b3-41db-4650-95bd-11df0dc06e80. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1176.686693] env[61987]: DEBUG nova.network.neutron [req-986d48a6-aaa5-4033-bd3f-09a01ce7d8f2 req-869ec86a-d4ad-46ec-8723-e78ca91246b7 service nova] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Updating instance_info_cache with network_info: [{"id": "79e1d3b3-41db-4650-95bd-11df0dc06e80", "address": "fa:16:3e:7f:09:fd", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79e1d3b3-41", "ovs_interfaceid": "79e1d3b3-41db-4650-95bd-11df0dc06e80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1176.704285] env[61987]: DEBUG oslo_concurrency.lockutils [None req-54794e69-6f33-470a-a452-f6d8ab941cc5 tempest-ServersTestBootFromVolume-1309788234 tempest-ServersTestBootFromVolume-1309788234-project-member] Lock "2ba77d32-d777-448e-8fa6-04920950f0f6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.549s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.753578] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523ee3f9-d8aa-cfa7-26dc-9d39b71beda7, 'name': SearchDatastore_Task, 'duration_secs': 0.012659} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.753921] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1176.754110] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1176.754359] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1176.754512] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.754696] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1176.755261] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-01f212b4-5de7-48fd-9452-28d5550a26ef {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.764574] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1176.764756] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1176.765503] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fceed768-d5a9-4ee0-ba91-0f83f378530b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.771458] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1176.771458] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]521595d6-65f4-f0f9-ae12-dc29828aab4b" [ 1176.771458] env[61987]: _type = "Task" [ 1176.771458] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.779087] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]521595d6-65f4-f0f9-ae12-dc29828aab4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.970151] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062467, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.189747] env[61987]: DEBUG oslo_concurrency.lockutils [req-986d48a6-aaa5-4033-bd3f-09a01ce7d8f2 req-869ec86a-d4ad-46ec-8723-e78ca91246b7 service nova] Releasing lock "refresh_cache-0704efaf-08de-4888-a092-476bcb789eea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1177.281870] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]521595d6-65f4-f0f9-ae12-dc29828aab4b, 'name': SearchDatastore_Task, 'duration_secs': 0.010372} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.282660] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a4d8b2e-092e-404a-b053-6d35f04913e2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.287705] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1177.287705] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52af07a3-5925-8238-85cf-308745cdb2a8" [ 1177.287705] env[61987]: _type = "Task" [ 1177.287705] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.295418] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52af07a3-5925-8238-85cf-308745cdb2a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.469798] env[61987]: DEBUG oslo_vmware.api [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062467, 'name': PowerOnVM_Task, 'duration_secs': 0.560175} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.470054] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1177.573797] env[61987]: DEBUG nova.compute.manager [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1177.574988] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-060eb055-246b-4b7d-bf7d-41753b47b7dc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.799866] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52af07a3-5925-8238-85cf-308745cdb2a8, 'name': SearchDatastore_Task, 'duration_secs': 0.009658} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.800148] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1177.800447] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 0704efaf-08de-4888-a092-476bcb789eea/0704efaf-08de-4888-a092-476bcb789eea.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1177.800635] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e363c1a8-c407-48eb-80e0-74818b166ba1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.807483] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1177.807483] env[61987]: value = "task-1062468" [ 1177.807483] env[61987]: _type = "Task" [ 1177.807483] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.815871] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062468, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.095986] env[61987]: DEBUG oslo_concurrency.lockutils [None req-e3e54c84-b3a6-458f-ae09-f6ca5257c0cd tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "f030044b-bc6b-478c-8f63-d519aa76203c" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 28.953s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.322174] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062468, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.821834] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062468, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.543395} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.822231] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore1] 0704efaf-08de-4888-a092-476bcb789eea/0704efaf-08de-4888-a092-476bcb789eea.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1178.822522] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1178.822875] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6ba4d23e-ec2a-4732-8828-29f3216a7318 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.832675] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1178.832675] env[61987]: value = "task-1062470" [ 1178.832675] env[61987]: _type = "Task" [ 1178.832675] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.847132] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062470, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.343700] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062470, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.354017} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.343919] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1179.344727] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a0d83c-466d-4a4f-93ed-a91d63592b14 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.367388] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Reconfiguring VM instance instance-00000075 to attach disk [datastore1] 0704efaf-08de-4888-a092-476bcb789eea/0704efaf-08de-4888-a092-476bcb789eea.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1179.367692] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0dde5b83-e559-45b6-823d-074788875d6a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.388640] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1179.388640] env[61987]: value = "task-1062471" [ 1179.388640] env[61987]: _type = "Task" [ 1179.388640] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.399117] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062471, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.898849] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062471, 'name': ReconfigVM_Task, 'duration_secs': 0.295808} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.899206] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Reconfigured VM instance instance-00000075 to attach disk [datastore1] 0704efaf-08de-4888-a092-476bcb789eea/0704efaf-08de-4888-a092-476bcb789eea.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1179.899930] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f512c8da-df6e-4c37-a65d-0e8da69c16d6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.907675] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1179.907675] env[61987]: value = "task-1062472" [ 1179.907675] env[61987]: _type = "Task" [ 1179.907675] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.918306] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062472, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.417628] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062472, 'name': Rename_Task, 'duration_secs': 0.150938} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.417944] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1180.418244] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8df0c179-4595-431a-8e78-88cc1f95e1db {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.424452] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1180.424452] env[61987]: value = "task-1062473" [ 1180.424452] env[61987]: _type = "Task" [ 1180.424452] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.432134] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062473, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.934828] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062473, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.437799] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062473, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.936865] env[61987]: DEBUG oslo_vmware.api [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062473, 'name': PowerOnVM_Task, 'duration_secs': 1.037385} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.936865] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1181.937231] env[61987]: INFO nova.compute.manager [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Took 8.37 seconds to spawn the instance on the hypervisor. [ 1181.937231] env[61987]: DEBUG nova.compute.manager [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1181.938023] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3103363a-57d3-4eaf-891e-944bb10c7169 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.099961] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "b64111ab-4cd4-4584-b64e-e61b129a721d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1182.100227] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "b64111ab-4cd4-4584-b64e-e61b129a721d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1182.455300] env[61987]: INFO nova.compute.manager [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Took 13.08 seconds to build instance. [ 1182.602705] env[61987]: DEBUG nova.compute.manager [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1182.717680] env[61987]: DEBUG nova.compute.manager [req-cb20dfdb-0800-4f58-959f-bb71c4549f2b req-1c18f953-14bd-4b44-8f6c-37f8f1c0faf5 service nova] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Received event network-changed-79e1d3b3-41db-4650-95bd-11df0dc06e80 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1182.717943] env[61987]: DEBUG nova.compute.manager [req-cb20dfdb-0800-4f58-959f-bb71c4549f2b req-1c18f953-14bd-4b44-8f6c-37f8f1c0faf5 service nova] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Refreshing instance network info cache due to event network-changed-79e1d3b3-41db-4650-95bd-11df0dc06e80. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1182.718231] env[61987]: DEBUG oslo_concurrency.lockutils [req-cb20dfdb-0800-4f58-959f-bb71c4549f2b req-1c18f953-14bd-4b44-8f6c-37f8f1c0faf5 service nova] Acquiring lock "refresh_cache-0704efaf-08de-4888-a092-476bcb789eea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1182.718439] env[61987]: DEBUG oslo_concurrency.lockutils [req-cb20dfdb-0800-4f58-959f-bb71c4549f2b req-1c18f953-14bd-4b44-8f6c-37f8f1c0faf5 service nova] Acquired lock "refresh_cache-0704efaf-08de-4888-a092-476bcb789eea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.718648] env[61987]: DEBUG nova.network.neutron [req-cb20dfdb-0800-4f58-959f-bb71c4549f2b req-1c18f953-14bd-4b44-8f6c-37f8f1c0faf5 service nova] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Refreshing network info cache for port 79e1d3b3-41db-4650-95bd-11df0dc06e80 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1182.957596] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5008965-f2eb-4a46-bf90-160795ab1d37 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "0704efaf-08de-4888-a092-476bcb789eea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.591s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.126591] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1183.126863] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1183.128465] env[61987]: INFO nova.compute.claims [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1183.422851] env[61987]: DEBUG nova.network.neutron [req-cb20dfdb-0800-4f58-959f-bb71c4549f2b req-1c18f953-14bd-4b44-8f6c-37f8f1c0faf5 service nova] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Updated VIF entry in instance network info cache for port 79e1d3b3-41db-4650-95bd-11df0dc06e80. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1183.423236] env[61987]: DEBUG nova.network.neutron [req-cb20dfdb-0800-4f58-959f-bb71c4549f2b req-1c18f953-14bd-4b44-8f6c-37f8f1c0faf5 service nova] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Updating instance_info_cache with network_info: [{"id": "79e1d3b3-41db-4650-95bd-11df0dc06e80", "address": "fa:16:3e:7f:09:fd", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79e1d3b3-41", "ovs_interfaceid": "79e1d3b3-41db-4650-95bd-11df0dc06e80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1183.925937] env[61987]: DEBUG oslo_concurrency.lockutils [req-cb20dfdb-0800-4f58-959f-bb71c4549f2b req-1c18f953-14bd-4b44-8f6c-37f8f1c0faf5 service nova] Releasing lock "refresh_cache-0704efaf-08de-4888-a092-476bcb789eea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1184.203818] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79ae016f-dda6-4a2f-a0f7-4244b2ee8bbe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.212204] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba9abd4-89c9-4b12-b68e-39cd32d35fa5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.243125] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36421d1d-b5c5-44d8-a8ec-f90208468d35 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.251203] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f81f29f8-4171-4b38-ae2a-98dabad753f5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.265089] env[61987]: DEBUG nova.compute.provider_tree [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1184.785347] env[61987]: ERROR nova.scheduler.client.report [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [req-01d151eb-5ea0-473c-b9e0-ce488bf727c0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8db744e5-1e14-4e98-9818-695ea4b7dfd0. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-01d151eb-5ea0-473c-b9e0-ce488bf727c0"}]} [ 1184.801896] env[61987]: DEBUG nova.scheduler.client.report [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Refreshing inventories for resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1184.818091] env[61987]: DEBUG nova.scheduler.client.report [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Updating ProviderTree inventory for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1184.818414] env[61987]: DEBUG nova.compute.provider_tree [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1184.829477] env[61987]: DEBUG nova.scheduler.client.report [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Refreshing aggregate associations for resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0, aggregates: None {{(pid=61987) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1184.846092] env[61987]: DEBUG nova.scheduler.client.report [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Refreshing trait associations for resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61987) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1184.906775] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c0114f9-5be8-4960-be3e-e2e671f8f742 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.914742] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd64488-3f5f-421f-aed0-caa52e67407b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.943281] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d07932-603d-4c7f-964d-a1fd2427d55b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.949944] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2356270a-6107-442e-91cd-e1d4afed773d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.963888] env[61987]: DEBUG nova.compute.provider_tree [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1185.495051] env[61987]: DEBUG nova.scheduler.client.report [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Updated inventory for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with generation 146 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1185.495051] env[61987]: DEBUG nova.compute.provider_tree [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Updating resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 generation from 146 to 147 during operation: update_inventory {{(pid=61987) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1185.495051] env[61987]: DEBUG nova.compute.provider_tree [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1186.000547] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.873s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1186.000964] env[61987]: DEBUG nova.compute.manager [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1186.506641] env[61987]: DEBUG nova.compute.utils [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1186.508382] env[61987]: DEBUG nova.compute.manager [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1186.508548] env[61987]: DEBUG nova.network.neutron [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1186.547674] env[61987]: DEBUG nova.policy [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c5cf7a7673cc4f5aadca7960c355c394', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ea5c65ac3b143178b13731663dbd8ee', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 1186.851641] env[61987]: DEBUG nova.network.neutron [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Successfully created port: d3a24d05-4d6d-4ae9-9f18-33c775b215ee {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1187.011789] env[61987]: DEBUG nova.compute.manager [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1187.517680] env[61987]: INFO nova.virt.block_device [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Booting with volume 6e8a12d3-a1ae-481c-ad5f-a05364a8dc87 at /dev/sda [ 1187.550326] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9083eb8e-ab9f-4384-83ca-c829ffd127a2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.560990] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93322ea8-54c5-4785-b46f-7838560e479b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.588078] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-275f4dab-c0fe-4317-8790-a871b83cdcad {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.596850] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6830adde-b9fd-4a31-a94b-9c5554ecd83b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.626290] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f56255f2-8ae9-4a44-a3e8-fb831c0aa695 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.633182] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b76fe2f-85e6-4b63-907c-40c85a042e53 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.646782] env[61987]: DEBUG nova.virt.block_device [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Updating existing volume attachment record: 56f32164-6d6a-4930-aabc-0884079c3943 {{(pid=61987) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1188.425925] env[61987]: DEBUG nova.compute.manager [req-dfb2e06a-f53d-4821-91af-f59bf3f04bd4 req-ef13c10d-c3d9-420d-a3c8-017ce718ae1c service nova] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Received event network-vif-plugged-d3a24d05-4d6d-4ae9-9f18-33c775b215ee {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1188.425925] env[61987]: DEBUG oslo_concurrency.lockutils [req-dfb2e06a-f53d-4821-91af-f59bf3f04bd4 req-ef13c10d-c3d9-420d-a3c8-017ce718ae1c service nova] Acquiring lock "b64111ab-4cd4-4584-b64e-e61b129a721d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1188.425925] env[61987]: DEBUG oslo_concurrency.lockutils [req-dfb2e06a-f53d-4821-91af-f59bf3f04bd4 req-ef13c10d-c3d9-420d-a3c8-017ce718ae1c service nova] Lock "b64111ab-4cd4-4584-b64e-e61b129a721d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1188.426117] env[61987]: DEBUG oslo_concurrency.lockutils [req-dfb2e06a-f53d-4821-91af-f59bf3f04bd4 req-ef13c10d-c3d9-420d-a3c8-017ce718ae1c service nova] Lock "b64111ab-4cd4-4584-b64e-e61b129a721d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1188.426349] env[61987]: DEBUG nova.compute.manager [req-dfb2e06a-f53d-4821-91af-f59bf3f04bd4 req-ef13c10d-c3d9-420d-a3c8-017ce718ae1c service nova] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] No waiting events found dispatching network-vif-plugged-d3a24d05-4d6d-4ae9-9f18-33c775b215ee {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1188.426478] env[61987]: WARNING nova.compute.manager [req-dfb2e06a-f53d-4821-91af-f59bf3f04bd4 req-ef13c10d-c3d9-420d-a3c8-017ce718ae1c service nova] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Received unexpected event network-vif-plugged-d3a24d05-4d6d-4ae9-9f18-33c775b215ee for instance with vm_state building and task_state block_device_mapping. [ 1188.509975] env[61987]: DEBUG nova.network.neutron [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Successfully updated port: d3a24d05-4d6d-4ae9-9f18-33c775b215ee {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1189.012885] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "refresh_cache-b64111ab-4cd4-4584-b64e-e61b129a721d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1189.013255] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired lock "refresh_cache-b64111ab-4cd4-4584-b64e-e61b129a721d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.013255] env[61987]: DEBUG nova.network.neutron [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1189.544602] env[61987]: DEBUG nova.network.neutron [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1189.673275] env[61987]: DEBUG nova.network.neutron [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Updating instance_info_cache with network_info: [{"id": "d3a24d05-4d6d-4ae9-9f18-33c775b215ee", "address": "fa:16:3e:d8:19:b8", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3a24d05-4d", "ovs_interfaceid": "d3a24d05-4d6d-4ae9-9f18-33c775b215ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1189.741370] env[61987]: DEBUG nova.compute.manager [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1189.741863] env[61987]: DEBUG nova.virt.hardware [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1189.742107] env[61987]: DEBUG nova.virt.hardware [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1189.742278] env[61987]: DEBUG nova.virt.hardware [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1189.742486] env[61987]: DEBUG nova.virt.hardware [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1189.742640] env[61987]: DEBUG nova.virt.hardware [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1189.742794] env[61987]: DEBUG nova.virt.hardware [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1189.742999] env[61987]: DEBUG nova.virt.hardware [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1189.743182] env[61987]: DEBUG nova.virt.hardware [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1189.743356] env[61987]: DEBUG nova.virt.hardware [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1189.743526] env[61987]: DEBUG nova.virt.hardware [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1189.743702] env[61987]: DEBUG nova.virt.hardware [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1189.744581] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8233730d-6de6-4f90-9420-7fc580299c77 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.754047] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-774c1ff3-dcbd-4b78-8a75-188ad1124f53 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.176029] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Releasing lock "refresh_cache-b64111ab-4cd4-4584-b64e-e61b129a721d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1190.176029] env[61987]: DEBUG nova.compute.manager [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Instance network_info: |[{"id": "d3a24d05-4d6d-4ae9-9f18-33c775b215ee", "address": "fa:16:3e:d8:19:b8", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3a24d05-4d", "ovs_interfaceid": "d3a24d05-4d6d-4ae9-9f18-33c775b215ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1190.176527] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:19:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'be8bd197-4b2b-46e7-88ea-2554b0438584', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd3a24d05-4d6d-4ae9-9f18-33c775b215ee', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1190.184137] env[61987]: DEBUG oslo.service.loopingcall [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1190.184368] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1190.184624] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9d2ae1f2-6172-4fa1-9cb5-c233692bfc20 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.206586] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1190.206586] env[61987]: value = "task-1062474" [ 1190.206586] env[61987]: _type = "Task" [ 1190.206586] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.214713] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062474, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.450342] env[61987]: DEBUG nova.compute.manager [req-9ceb11c8-3b91-42d7-b268-0fb216214d31 req-6663903f-2492-4d86-a18e-ac00ee640da7 service nova] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Received event network-changed-d3a24d05-4d6d-4ae9-9f18-33c775b215ee {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1190.450439] env[61987]: DEBUG nova.compute.manager [req-9ceb11c8-3b91-42d7-b268-0fb216214d31 req-6663903f-2492-4d86-a18e-ac00ee640da7 service nova] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Refreshing instance network info cache due to event network-changed-d3a24d05-4d6d-4ae9-9f18-33c775b215ee. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1190.450736] env[61987]: DEBUG oslo_concurrency.lockutils [req-9ceb11c8-3b91-42d7-b268-0fb216214d31 req-6663903f-2492-4d86-a18e-ac00ee640da7 service nova] Acquiring lock "refresh_cache-b64111ab-4cd4-4584-b64e-e61b129a721d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1190.450931] env[61987]: DEBUG oslo_concurrency.lockutils [req-9ceb11c8-3b91-42d7-b268-0fb216214d31 req-6663903f-2492-4d86-a18e-ac00ee640da7 service nova] Acquired lock "refresh_cache-b64111ab-4cd4-4584-b64e-e61b129a721d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1190.451140] env[61987]: DEBUG nova.network.neutron [req-9ceb11c8-3b91-42d7-b268-0fb216214d31 req-6663903f-2492-4d86-a18e-ac00ee640da7 service nova] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Refreshing network info cache for port d3a24d05-4d6d-4ae9-9f18-33c775b215ee {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1190.717361] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062474, 'name': CreateVM_Task, 'duration_secs': 0.318385} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.717577] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1190.718138] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'device_type': None, 'attachment_id': '56f32164-6d6a-4930-aabc-0884079c3943', 'mount_device': '/dev/sda', 'delete_on_termination': True, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234455', 'volume_id': '6e8a12d3-a1ae-481c-ad5f-a05364a8dc87', 'name': 'volume-6e8a12d3-a1ae-481c-ad5f-a05364a8dc87', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b64111ab-4cd4-4584-b64e-e61b129a721d', 'attached_at': '', 'detached_at': '', 'volume_id': '6e8a12d3-a1ae-481c-ad5f-a05364a8dc87', 'serial': '6e8a12d3-a1ae-481c-ad5f-a05364a8dc87'}, 'boot_index': 0, 'guest_format': None, 'disk_bus': None, 'volume_type': None}], 'swap': None} {{(pid=61987) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1190.718438] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Root volume attach. Driver type: vmdk {{(pid=61987) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1190.719204] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93570638-644a-4217-997d-ddb163ef06a7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.726985] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b36cddfb-7e41-470f-94c3-af432141ded6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.732730] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5034075e-bef0-4d1c-bd08-2adfcdfb6d90 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.738439] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-9914bbc0-4573-4242-91a9-85e26bd7cc7b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.746047] env[61987]: DEBUG oslo_vmware.api [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1190.746047] env[61987]: value = "task-1062475" [ 1190.746047] env[61987]: _type = "Task" [ 1190.746047] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.753643] env[61987]: DEBUG oslo_vmware.api [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062475, 'name': RelocateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.167252] env[61987]: DEBUG nova.network.neutron [req-9ceb11c8-3b91-42d7-b268-0fb216214d31 req-6663903f-2492-4d86-a18e-ac00ee640da7 service nova] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Updated VIF entry in instance network info cache for port d3a24d05-4d6d-4ae9-9f18-33c775b215ee. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1191.167677] env[61987]: DEBUG nova.network.neutron [req-9ceb11c8-3b91-42d7-b268-0fb216214d31 req-6663903f-2492-4d86-a18e-ac00ee640da7 service nova] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Updating instance_info_cache with network_info: [{"id": "d3a24d05-4d6d-4ae9-9f18-33c775b215ee", "address": "fa:16:3e:d8:19:b8", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3a24d05-4d", "ovs_interfaceid": "d3a24d05-4d6d-4ae9-9f18-33c775b215ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1191.260488] env[61987]: DEBUG oslo_vmware.api [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062475, 'name': RelocateVM_Task} progress is 42%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.670504] env[61987]: DEBUG oslo_concurrency.lockutils [req-9ceb11c8-3b91-42d7-b268-0fb216214d31 req-6663903f-2492-4d86-a18e-ac00ee640da7 service nova] Releasing lock "refresh_cache-b64111ab-4cd4-4584-b64e-e61b129a721d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1191.761666] env[61987]: DEBUG oslo_vmware.api [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062475, 'name': RelocateVM_Task} progress is 56%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.263695] env[61987]: DEBUG oslo_vmware.api [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062475, 'name': RelocateVM_Task} progress is 71%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.764192] env[61987]: DEBUG oslo_vmware.api [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062475, 'name': RelocateVM_Task} progress is 84%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.259597] env[61987]: DEBUG oslo_vmware.api [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062475, 'name': RelocateVM_Task} progress is 97%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.759940] env[61987]: DEBUG oslo_vmware.api [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062475, 'name': RelocateVM_Task} progress is 98%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.260610] env[61987]: DEBUG oslo_vmware.api [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062475, 'name': RelocateVM_Task, 'duration_secs': 3.07304} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.260925] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Volume attach. Driver type: vmdk {{(pid=61987) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1194.261152] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234455', 'volume_id': '6e8a12d3-a1ae-481c-ad5f-a05364a8dc87', 'name': 'volume-6e8a12d3-a1ae-481c-ad5f-a05364a8dc87', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b64111ab-4cd4-4584-b64e-e61b129a721d', 'attached_at': '', 'detached_at': '', 'volume_id': '6e8a12d3-a1ae-481c-ad5f-a05364a8dc87', 'serial': '6e8a12d3-a1ae-481c-ad5f-a05364a8dc87'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1194.261914] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-078bce13-d838-4637-8a6c-2dae718a0917 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.277046] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8340b3b-518b-4c12-9394-ff5984fe995f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.302460] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Reconfiguring VM instance instance-00000076 to attach disk [datastore2] volume-6e8a12d3-a1ae-481c-ad5f-a05364a8dc87/volume-6e8a12d3-a1ae-481c-ad5f-a05364a8dc87.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1194.302827] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1747cb95-5dd5-4fd8-a962-af3dc54e4785 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.331806] env[61987]: DEBUG oslo_vmware.api [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1194.331806] env[61987]: value = "task-1062476" [ 1194.331806] env[61987]: _type = "Task" [ 1194.331806] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.340905] env[61987]: DEBUG oslo_vmware.api [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062476, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.841986] env[61987]: DEBUG oslo_vmware.api [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062476, 'name': ReconfigVM_Task, 'duration_secs': 0.253018} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1194.842376] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Reconfigured VM instance instance-00000076 to attach disk [datastore2] volume-6e8a12d3-a1ae-481c-ad5f-a05364a8dc87/volume-6e8a12d3-a1ae-481c-ad5f-a05364a8dc87.vmdk or device None with type thin {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1194.847128] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-388ff634-0892-4eaf-bc8e-062dfd1938bf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.862243] env[61987]: DEBUG oslo_vmware.api [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1194.862243] env[61987]: value = "task-1062477" [ 1194.862243] env[61987]: _type = "Task" [ 1194.862243] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.870068] env[61987]: DEBUG oslo_vmware.api [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062477, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.372718] env[61987]: DEBUG oslo_vmware.api [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062477, 'name': ReconfigVM_Task, 'duration_secs': 0.130529} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.373038] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234455', 'volume_id': '6e8a12d3-a1ae-481c-ad5f-a05364a8dc87', 'name': 'volume-6e8a12d3-a1ae-481c-ad5f-a05364a8dc87', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b64111ab-4cd4-4584-b64e-e61b129a721d', 'attached_at': '', 'detached_at': '', 'volume_id': '6e8a12d3-a1ae-481c-ad5f-a05364a8dc87', 'serial': '6e8a12d3-a1ae-481c-ad5f-a05364a8dc87'} {{(pid=61987) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1195.373568] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b3899a37-18e2-46b7-a3fb-457c652967fc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.380451] env[61987]: DEBUG oslo_vmware.api [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1195.380451] env[61987]: value = "task-1062478" [ 1195.380451] env[61987]: _type = "Task" [ 1195.380451] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.387922] env[61987]: DEBUG oslo_vmware.api [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062478, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.890195] env[61987]: DEBUG oslo_vmware.api [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062478, 'name': Rename_Task, 'duration_secs': 0.126706} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.890557] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1195.890714] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6077defd-f816-407b-9ed5-6a8b72272af8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.897645] env[61987]: DEBUG oslo_vmware.api [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1195.897645] env[61987]: value = "task-1062479" [ 1195.897645] env[61987]: _type = "Task" [ 1195.897645] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.907372] env[61987]: DEBUG oslo_vmware.api [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062479, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.407307] env[61987]: DEBUG oslo_vmware.api [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062479, 'name': PowerOnVM_Task, 'duration_secs': 0.42675} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.407589] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1196.407797] env[61987]: INFO nova.compute.manager [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Took 6.67 seconds to spawn the instance on the hypervisor. [ 1196.407982] env[61987]: DEBUG nova.compute.manager [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1196.408786] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce8ef1c8-37a9-4896-9057-9f588ee58269 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.928687] env[61987]: INFO nova.compute.manager [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Took 13.82 seconds to build instance. [ 1196.964806] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.965022] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.965174] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61987) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11020}} [ 1197.430957] env[61987]: DEBUG oslo_concurrency.lockutils [None req-9242caa5-3089-4fc1-bd11-d6c3f0809b8f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "b64111ab-4cd4-4584-b64e-e61b129a721d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.331s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1197.614277] env[61987]: DEBUG nova.compute.manager [req-87a1fce4-1c16-4aaa-abf9-57a013e99e9b req-ab761a95-aab3-4388-8521-6fe1bd175dd7 service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Received event network-changed-203d81e4-dbb8-4685-a463-9de48b835ff7 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1197.614574] env[61987]: DEBUG nova.compute.manager [req-87a1fce4-1c16-4aaa-abf9-57a013e99e9b req-ab761a95-aab3-4388-8521-6fe1bd175dd7 service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Refreshing instance network info cache due to event network-changed-203d81e4-dbb8-4685-a463-9de48b835ff7. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1197.614777] env[61987]: DEBUG oslo_concurrency.lockutils [req-87a1fce4-1c16-4aaa-abf9-57a013e99e9b req-ab761a95-aab3-4388-8521-6fe1bd175dd7 service nova] Acquiring lock "refresh_cache-4d15848d-5265-4897-bfa7-f3ad4c02ed83" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1197.615811] env[61987]: DEBUG oslo_concurrency.lockutils [req-87a1fce4-1c16-4aaa-abf9-57a013e99e9b req-ab761a95-aab3-4388-8521-6fe1bd175dd7 service nova] Acquired lock "refresh_cache-4d15848d-5265-4897-bfa7-f3ad4c02ed83" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1197.615811] env[61987]: DEBUG nova.network.neutron [req-87a1fce4-1c16-4aaa-abf9-57a013e99e9b req-ab761a95-aab3-4388-8521-6fe1bd175dd7 service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Refreshing network info cache for port 203d81e4-dbb8-4685-a463-9de48b835ff7 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1197.965175] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1197.965452] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Starting heal instance info cache {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10401}} [ 1197.965496] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Rebuilding the list of instances to heal {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10405}} [ 1198.312994] env[61987]: DEBUG nova.network.neutron [req-87a1fce4-1c16-4aaa-abf9-57a013e99e9b req-ab761a95-aab3-4388-8521-6fe1bd175dd7 service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Updated VIF entry in instance network info cache for port 203d81e4-dbb8-4685-a463-9de48b835ff7. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1198.313409] env[61987]: DEBUG nova.network.neutron [req-87a1fce4-1c16-4aaa-abf9-57a013e99e9b req-ab761a95-aab3-4388-8521-6fe1bd175dd7 service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Updating instance_info_cache with network_info: [{"id": "203d81e4-dbb8-4685-a463-9de48b835ff7", "address": "fa:16:3e:5c:c5:2c", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap203d81e4-db", "ovs_interfaceid": "203d81e4-dbb8-4685-a463-9de48b835ff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1198.497350] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "refresh_cache-4d15848d-5265-4897-bfa7-f3ad4c02ed83" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1198.816273] env[61987]: DEBUG oslo_concurrency.lockutils [req-87a1fce4-1c16-4aaa-abf9-57a013e99e9b req-ab761a95-aab3-4388-8521-6fe1bd175dd7 service nova] Releasing lock "refresh_cache-4d15848d-5265-4897-bfa7-f3ad4c02ed83" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1198.816676] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquired lock "refresh_cache-4d15848d-5265-4897-bfa7-f3ad4c02ed83" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1198.816838] env[61987]: DEBUG nova.network.neutron [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Forcefully refreshing network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1198.816997] env[61987]: DEBUG nova.objects.instance [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lazy-loading 'info_cache' on Instance uuid 4d15848d-5265-4897-bfa7-f3ad4c02ed83 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1199.041093] env[61987]: DEBUG nova.compute.manager [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Stashing vm_state: active {{(pid=61987) _prep_resize /opt/stack/nova/nova/compute/manager.py:6090}} [ 1199.562036] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1199.562036] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1199.637445] env[61987]: DEBUG nova.compute.manager [req-b5b68212-e891-4daf-8fa8-3f8d2b4d2988 req-9a34f5c5-85d4-4128-ab23-750b1cbce7da service nova] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Received event network-changed-d3a24d05-4d6d-4ae9-9f18-33c775b215ee {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1199.637655] env[61987]: DEBUG nova.compute.manager [req-b5b68212-e891-4daf-8fa8-3f8d2b4d2988 req-9a34f5c5-85d4-4128-ab23-750b1cbce7da service nova] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Refreshing instance network info cache due to event network-changed-d3a24d05-4d6d-4ae9-9f18-33c775b215ee. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1199.638246] env[61987]: DEBUG oslo_concurrency.lockutils [req-b5b68212-e891-4daf-8fa8-3f8d2b4d2988 req-9a34f5c5-85d4-4128-ab23-750b1cbce7da service nova] Acquiring lock "refresh_cache-b64111ab-4cd4-4584-b64e-e61b129a721d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1199.638440] env[61987]: DEBUG oslo_concurrency.lockutils [req-b5b68212-e891-4daf-8fa8-3f8d2b4d2988 req-9a34f5c5-85d4-4128-ab23-750b1cbce7da service nova] Acquired lock "refresh_cache-b64111ab-4cd4-4584-b64e-e61b129a721d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.638620] env[61987]: DEBUG nova.network.neutron [req-b5b68212-e891-4daf-8fa8-3f8d2b4d2988 req-9a34f5c5-85d4-4128-ab23-750b1cbce7da service nova] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Refreshing network info cache for port d3a24d05-4d6d-4ae9-9f18-33c775b215ee {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1200.066916] env[61987]: INFO nova.compute.claims [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1200.350664] env[61987]: DEBUG nova.network.neutron [req-b5b68212-e891-4daf-8fa8-3f8d2b4d2988 req-9a34f5c5-85d4-4128-ab23-750b1cbce7da service nova] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Updated VIF entry in instance network info cache for port d3a24d05-4d6d-4ae9-9f18-33c775b215ee. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1200.351091] env[61987]: DEBUG nova.network.neutron [req-b5b68212-e891-4daf-8fa8-3f8d2b4d2988 req-9a34f5c5-85d4-4128-ab23-750b1cbce7da service nova] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Updating instance_info_cache with network_info: [{"id": "d3a24d05-4d6d-4ae9-9f18-33c775b215ee", "address": "fa:16:3e:d8:19:b8", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3a24d05-4d", "ovs_interfaceid": "d3a24d05-4d6d-4ae9-9f18-33c775b215ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1200.505734] env[61987]: DEBUG nova.network.neutron [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Updating instance_info_cache with network_info: [{"id": "203d81e4-dbb8-4685-a463-9de48b835ff7", "address": "fa:16:3e:5c:c5:2c", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap203d81e4-db", "ovs_interfaceid": "203d81e4-dbb8-4685-a463-9de48b835ff7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1200.573054] env[61987]: INFO nova.compute.resource_tracker [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Updating resource usage from migration dfc37271-97fb-4dfa-9ce4-0aee9b61adc4 [ 1200.645864] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e24fb084-bfb4-403b-b5b7-10d172ed5b46 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.654060] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-460020bc-0b17-439c-9b13-99abe5fa0305 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.685489] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5131711e-6b50-4086-b1d8-c421f2777cf8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.693180] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd16d780-4a8a-436d-ad5d-e0be0c858150 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.706160] env[61987]: DEBUG nova.compute.provider_tree [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1200.853851] env[61987]: DEBUG oslo_concurrency.lockutils [req-b5b68212-e891-4daf-8fa8-3f8d2b4d2988 req-9a34f5c5-85d4-4128-ab23-750b1cbce7da service nova] Releasing lock "refresh_cache-b64111ab-4cd4-4584-b64e-e61b129a721d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1201.008922] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Releasing lock "refresh_cache-4d15848d-5265-4897-bfa7-f3ad4c02ed83" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1201.009102] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Updated the network info_cache for instance {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10472}} [ 1201.009292] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1201.009482] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1201.009650] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1201.009803] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1201.224859] env[61987]: ERROR nova.scheduler.client.report [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [req-48d151b4-8bb5-4f43-b6bb-b0e98f41c56b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8db744e5-1e14-4e98-9818-695ea4b7dfd0. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-48d151b4-8bb5-4f43-b6bb-b0e98f41c56b"}]} [ 1201.240597] env[61987]: DEBUG nova.scheduler.client.report [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Refreshing inventories for resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1201.253955] env[61987]: DEBUG nova.scheduler.client.report [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Updating ProviderTree inventory for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1201.254203] env[61987]: DEBUG nova.compute.provider_tree [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1201.267538] env[61987]: DEBUG nova.scheduler.client.report [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Refreshing aggregate associations for resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0, aggregates: None {{(pid=61987) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1201.284813] env[61987]: DEBUG nova.scheduler.client.report [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Refreshing trait associations for resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61987) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1201.356354] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9369e10-9803-46d3-a0be-aa218ddfe94c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.364787] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a4f9c17-b38c-4e09-8756-684f782c1525 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.395281] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9143e2ba-5669-4556-b646-ebf5e9da48d6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.402386] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c827af9c-09e2-4042-a684-665bbaaf7a22 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.415363] env[61987]: DEBUG nova.compute.provider_tree [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1201.513049] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1201.945156] env[61987]: DEBUG nova.scheduler.client.report [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Updated inventory for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with generation 149 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1201.945441] env[61987]: DEBUG nova.compute.provider_tree [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Updating resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 generation from 149 to 150 during operation: update_inventory {{(pid=61987) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1201.945629] env[61987]: DEBUG nova.compute.provider_tree [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1202.450685] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.889s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.450998] env[61987]: INFO nova.compute.manager [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Migrating [ 1202.457660] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.945s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1202.457880] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1202.458082] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61987) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1202.461381] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-711958fb-15f7-4320-a9db-36a242f16a60 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.474887] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2167554d-e9df-41de-8f1d-b1e4f9ab6345 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.490663] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7474d0c8-9c84-4dde-ade4-f9ca659109bc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.498200] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c3537f-349d-492f-bba9-4e20e005b80a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.529586] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181058MB free_disk=179GB free_vcpus=48 pci_devices=None {{(pid=61987) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1202.529781] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1202.530041] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1202.969756] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "refresh_cache-b64111ab-4cd4-4584-b64e-e61b129a721d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1202.970039] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquired lock "refresh_cache-b64111ab-4cd4-4584-b64e-e61b129a721d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1202.970166] env[61987]: DEBUG nova.network.neutron [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1203.538460] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Applying migration context for instance b64111ab-4cd4-4584-b64e-e61b129a721d as it has an incoming, in-progress migration dfc37271-97fb-4dfa-9ce4-0aee9b61adc4. Migration status is pre-migrating {{(pid=61987) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1203.539371] env[61987]: INFO nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Updating resource usage from migration dfc37271-97fb-4dfa-9ce4-0aee9b61adc4 [ 1203.557022] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 4d15848d-5265-4897-bfa7-f3ad4c02ed83 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1203.557022] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1203.557022] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance f030044b-bc6b-478c-8f63-d519aa76203c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1203.557022] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 0704efaf-08de-4888-a092-476bcb789eea actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1203.557022] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Migration dfc37271-97fb-4dfa-9ce4-0aee9b61adc4 is active on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1203.557022] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance b64111ab-4cd4-4584-b64e-e61b129a721d actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1203.557301] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1203.557301] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1728MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1203.637693] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cb6734f-96bd-4ae7-9c72-00ff9764c204 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.645190] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13945b07-f249-477e-b908-25fb029c2909 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.676137] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64670282-ce38-4805-b305-a00516a29c89 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.684015] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f101e3-5c78-431c-914e-e7296c02710c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.698397] env[61987]: DEBUG nova.compute.provider_tree [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1203.713818] env[61987]: DEBUG nova.network.neutron [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Updating instance_info_cache with network_info: [{"id": "d3a24d05-4d6d-4ae9-9f18-33c775b215ee", "address": "fa:16:3e:d8:19:b8", "network": {"id": "924c2b1d-33c2-4692-aee0-42e75e6367c4", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-2129231470-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ea5c65ac3b143178b13731663dbd8ee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "be8bd197-4b2b-46e7-88ea-2554b0438584", "external-id": "nsx-vlan-transportzone-338", "segmentation_id": 338, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3a24d05-4d", "ovs_interfaceid": "d3a24d05-4d6d-4ae9-9f18-33c775b215ee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1204.201783] env[61987]: DEBUG nova.scheduler.client.report [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1204.215676] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Releasing lock "refresh_cache-b64111ab-4cd4-4584-b64e-e61b129a721d" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1204.707026] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61987) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1204.707420] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.177s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1205.663700] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1205.664033] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1205.730288] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51db7eee-0e49-485c-b2b5-e4b63356acc1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.749937] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Updating instance 'b64111ab-4cd4-4584-b64e-e61b129a721d' progress to 0 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1206.256685] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1206.257037] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5758e75d-20bf-4dbd-b051-b2e0f51033c7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.264506] env[61987]: DEBUG oslo_vmware.api [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1206.264506] env[61987]: value = "task-1062480" [ 1206.264506] env[61987]: _type = "Task" [ 1206.264506] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.272957] env[61987]: DEBUG oslo_vmware.api [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062480, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.774666] env[61987]: DEBUG oslo_vmware.api [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062480, 'name': PowerOffVM_Task, 'duration_secs': 0.207706} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.774666] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1206.774666] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Updating instance 'b64111ab-4cd4-4584-b64e-e61b129a721d' progress to 17 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1207.281109] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:47Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1207.281387] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1207.281531] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1207.281728] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1207.281883] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1207.282051] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1207.282267] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1207.282432] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1207.282604] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1207.282775] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1207.282958] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1207.288078] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a4c5769a-db60-40cf-8377-d794f0969458 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.304626] env[61987]: DEBUG oslo_vmware.api [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1207.304626] env[61987]: value = "task-1062481" [ 1207.304626] env[61987]: _type = "Task" [ 1207.304626] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.312663] env[61987]: DEBUG oslo_vmware.api [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062481, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.814260] env[61987]: DEBUG oslo_vmware.api [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062481, 'name': ReconfigVM_Task, 'duration_secs': 0.152085} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.814633] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Updating instance 'b64111ab-4cd4-4584-b64e-e61b129a721d' progress to 33 {{(pid=61987) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1208.321309] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1208.321623] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1208.321750] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1208.321945] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1208.322114] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1208.322271] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1208.322475] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1208.322643] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1208.322816] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1208.322986] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1208.323192] env[61987]: DEBUG nova.virt.hardware [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1208.831235] env[61987]: ERROR nova.compute.manager [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Setting instance vm_state to ERROR: AttributeError: 'NoneType' object has no attribute 'key' [ 1208.831235] env[61987]: ERROR nova.compute.manager [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Traceback (most recent call last): [ 1208.831235] env[61987]: ERROR nova.compute.manager [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] File "/opt/stack/nova/nova/compute/manager.py", line 11338, in _error_out_instance_on_exception [ 1208.831235] env[61987]: ERROR nova.compute.manager [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] yield [ 1208.831235] env[61987]: ERROR nova.compute.manager [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] File "/opt/stack/nova/nova/compute/manager.py", line 6571, in _resize_instance [ 1208.831235] env[61987]: ERROR nova.compute.manager [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] disk_info = self.driver.migrate_disk_and_power_off( [ 1208.831235] env[61987]: ERROR nova.compute.manager [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 266, in migrate_disk_and_power_off [ 1208.831235] env[61987]: ERROR nova.compute.manager [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] return self._vmops.migrate_disk_and_power_off(context, instance, [ 1208.831235] env[61987]: ERROR nova.compute.manager [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1208.831235] env[61987]: ERROR nova.compute.manager [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1208.831235] env[61987]: ERROR nova.compute.manager [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1208.831235] env[61987]: ERROR nova.compute.manager [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1208.831235] env[61987]: ERROR nova.compute.manager [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1208.831235] env[61987]: ERROR nova.compute.manager [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] disk_key = device.key [ 1208.831235] env[61987]: ERROR nova.compute.manager [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] AttributeError: 'NoneType' object has no attribute 'key' [ 1208.831235] env[61987]: ERROR nova.compute.manager [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] [ 1209.348572] env[61987]: INFO nova.compute.manager [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Swapping old allocation on dict_keys(['8db744e5-1e14-4e98-9818-695ea4b7dfd0']) held by migration dfc37271-97fb-4dfa-9ce4-0aee9b61adc4 for instance [ 1209.369162] env[61987]: DEBUG nova.scheduler.client.report [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Overwriting current allocation {'allocations': {'8db744e5-1e14-4e98-9818-695ea4b7dfd0': {'resources': {'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 150}}, 'project_id': '8ea5c65ac3b143178b13731663dbd8ee', 'user_id': 'c5cf7a7673cc4f5aadca7960c355c394', 'consumer_generation': 1} on consumer b64111ab-4cd4-4584-b64e-e61b129a721d {{(pid=61987) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 1209.769074] env[61987]: DEBUG oslo_concurrency.lockutils [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "b64111ab-4cd4-4584-b64e-e61b129a721d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1209.769290] env[61987]: DEBUG oslo_concurrency.lockutils [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "b64111ab-4cd4-4584-b64e-e61b129a721d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1209.769431] env[61987]: DEBUG oslo_concurrency.lockutils [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "b64111ab-4cd4-4584-b64e-e61b129a721d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1209.769609] env[61987]: DEBUG oslo_concurrency.lockutils [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "b64111ab-4cd4-4584-b64e-e61b129a721d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1209.769809] env[61987]: DEBUG oslo_concurrency.lockutils [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "b64111ab-4cd4-4584-b64e-e61b129a721d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1209.771864] env[61987]: INFO nova.compute.manager [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Terminating instance [ 1209.961067] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1210.275908] env[61987]: DEBUG nova.compute.manager [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1210.276184] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1210.276485] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e1f9b6a0-a951-4058-9c02-c4636af7edfb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.283613] env[61987]: DEBUG oslo_vmware.api [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1210.283613] env[61987]: value = "task-1062482" [ 1210.283613] env[61987]: _type = "Task" [ 1210.283613] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.291095] env[61987]: DEBUG oslo_vmware.api [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062482, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.793924] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] VM already powered off {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1210.794226] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Volume detach. Driver type: vmdk {{(pid=61987) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1210.794441] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234455', 'volume_id': '6e8a12d3-a1ae-481c-ad5f-a05364a8dc87', 'name': 'volume-6e8a12d3-a1ae-481c-ad5f-a05364a8dc87', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b64111ab-4cd4-4584-b64e-e61b129a721d', 'attached_at': '', 'detached_at': '', 'volume_id': '6e8a12d3-a1ae-481c-ad5f-a05364a8dc87', 'serial': '6e8a12d3-a1ae-481c-ad5f-a05364a8dc87'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1210.795197] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d619522-3849-42ee-89c7-ed3a81b8ae50 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.813244] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dce77fa3-f356-4f68-9078-f9445ae21073 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.819207] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fadeecf4-ba8b-45d6-8aac-e71ef79c4752 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.836141] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a137e0-a59e-4a2c-894c-5135195fb0fd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.849660] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] The volume has not been displaced from its original location: [datastore2] volume-6e8a12d3-a1ae-481c-ad5f-a05364a8dc87/volume-6e8a12d3-a1ae-481c-ad5f-a05364a8dc87.vmdk. No consolidation needed. {{(pid=61987) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1210.854729] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Reconfiguring VM instance instance-00000076 to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1210.854975] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3097b3e8-e0b1-47f4-926d-11eb90061e89 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.871201] env[61987]: DEBUG oslo_vmware.api [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1210.871201] env[61987]: value = "task-1062483" [ 1210.871201] env[61987]: _type = "Task" [ 1210.871201] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.878011] env[61987]: DEBUG oslo_vmware.api [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062483, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.921307] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1210.921554] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.990120] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1974b9f5-a992-4a69-b238-306e9bc4c8e4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.000744] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-620f74a7-fa3f-4c21-9c99-ccab07f68d30 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.031504] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be597843-c340-41f4-9c5c-aa6730e202a4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.038065] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18073e0b-c265-469d-926e-c5df16834843 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.050813] env[61987]: DEBUG nova.compute.provider_tree [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1211.067505] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2bd01824-9c8b-494a-b29d-a691c2f784a2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "b64111ab-4cd4-4584-b64e-e61b129a721d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1211.381607] env[61987]: DEBUG oslo_vmware.api [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062483, 'name': ReconfigVM_Task, 'duration_secs': 0.157706} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.381852] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Reconfigured VM instance instance-00000076 to detach disk 2000 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1211.386410] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-84932a2a-c815-41e8-8ca6-7b5089ecb071 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.400195] env[61987]: DEBUG oslo_vmware.api [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1211.400195] env[61987]: value = "task-1062484" [ 1211.400195] env[61987]: _type = "Task" [ 1211.400195] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.407511] env[61987]: DEBUG oslo_vmware.api [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062484, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.553989] env[61987]: DEBUG nova.scheduler.client.report [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1211.909896] env[61987]: DEBUG oslo_vmware.api [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062484, 'name': ReconfigVM_Task, 'duration_secs': 0.101086} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.910224] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234455', 'volume_id': '6e8a12d3-a1ae-481c-ad5f-a05364a8dc87', 'name': 'volume-6e8a12d3-a1ae-481c-ad5f-a05364a8dc87', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b64111ab-4cd4-4584-b64e-e61b129a721d', 'attached_at': '', 'detached_at': '', 'volume_id': '6e8a12d3-a1ae-481c-ad5f-a05364a8dc87', 'serial': '6e8a12d3-a1ae-481c-ad5f-a05364a8dc87'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1211.910501] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1211.911288] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63539fd0-f8d8-430e-baab-68bc4620eeaa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.917340] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1211.917526] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e51d570b-6da9-42fc-80c2-d50ead1fe31f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.986787] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1211.986991] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1211.987208] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Deleting the datastore file [datastore2] b64111ab-4cd4-4584-b64e-e61b129a721d {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1211.987468] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9f1d71dd-8556-4588-a558-5b3f202264a6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.994056] env[61987]: DEBUG oslo_vmware.api [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1211.994056] env[61987]: value = "task-1062486" [ 1211.994056] env[61987]: _type = "Task" [ 1211.994056] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.001870] env[61987]: DEBUG oslo_vmware.api [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062486, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.059416] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.138s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1212.059715] env[61987]: INFO nova.compute.manager [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Successfully reverted task state from resize_migrating on failure for instance. [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server [None req-ef2d552e-d61b-4f63-b8cf-52013fc7cb04 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Exception during message handling: AttributeError: 'NoneType' object has no attribute 'key' [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 172, in _process_incoming [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server raise self.value [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 168, in decorated_function [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server raise self.value [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 159, in decorated_function [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1484, in decorated_function [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 215, in decorated_function [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server raise self.value [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 205, in decorated_function [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6537, in resize_instance [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server raise self.value [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6534, in resize_instance [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server self._resize_instance(context, instance, image, migration, [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6571, in _resize_instance [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server disk_info = self.driver.migrate_disk_and_power_off( [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 266, in migrate_disk_and_power_off [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server return self._vmops.migrate_disk_and_power_off(context, instance, [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server disk_key = device.key [ 1212.068682] env[61987]: ERROR oslo_messaging.rpc.server AttributeError: 'NoneType' object has no attribute 'key' [ 1212.070817] env[61987]: ERROR oslo_messaging.rpc.server [ 1212.368350] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.368690] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1212.368916] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1212.369130] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1212.369312] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1212.371547] env[61987]: INFO nova.compute.manager [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Terminating instance [ 1212.503425] env[61987]: DEBUG oslo_vmware.api [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062486, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.075406} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.503621] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1212.503813] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1212.503994] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1212.504190] env[61987]: INFO nova.compute.manager [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Took 2.23 seconds to destroy the instance on the hypervisor. [ 1212.504430] env[61987]: DEBUG oslo.service.loopingcall [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1212.504627] env[61987]: DEBUG nova.compute.manager [-] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1212.504724] env[61987]: DEBUG nova.network.neutron [-] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1212.875781] env[61987]: DEBUG nova.compute.manager [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1212.876063] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1212.877013] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c8141b1-9b56-4082-9c3d-ca16891126e4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.885995] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1212.886257] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d3444c0-eebd-4f5d-9d15-04b3282ece7f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.892387] env[61987]: DEBUG oslo_vmware.api [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1212.892387] env[61987]: value = "task-1062487" [ 1212.892387] env[61987]: _type = "Task" [ 1212.892387] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.904120] env[61987]: DEBUG oslo_vmware.api [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062487, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.970808] env[61987]: DEBUG nova.compute.manager [req-9728c917-477a-40dd-bd79-34560324f44c req-d6100dc3-0dfc-4833-a3fa-1935a6eb965f service nova] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Received event network-vif-deleted-d3a24d05-4d6d-4ae9-9f18-33c775b215ee {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1212.971966] env[61987]: INFO nova.compute.manager [req-9728c917-477a-40dd-bd79-34560324f44c req-d6100dc3-0dfc-4833-a3fa-1935a6eb965f service nova] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Neutron deleted interface d3a24d05-4d6d-4ae9-9f18-33c775b215ee; detaching it from the instance and deleting it from the info cache [ 1212.971966] env[61987]: DEBUG nova.network.neutron [req-9728c917-477a-40dd-bd79-34560324f44c req-d6100dc3-0dfc-4833-a3fa-1935a6eb965f service nova] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1213.402377] env[61987]: DEBUG oslo_vmware.api [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062487, 'name': PowerOffVM_Task, 'duration_secs': 0.347486} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.402763] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1213.402814] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1213.403084] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-17fd462e-a778-4b67-b1dc-aff54682e31f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.437130] env[61987]: DEBUG nova.network.neutron [-] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1213.470028] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1213.470312] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1213.470551] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Deleting the datastore file [datastore1] 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1213.470823] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-977f8c61-ff9f-44fe-9ef8-0d2bba16641a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.473746] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bc455967-32af-4db7-b117-e731dd9bd9c4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.476646] env[61987]: DEBUG oslo_vmware.api [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1213.476646] env[61987]: value = "task-1062489" [ 1213.476646] env[61987]: _type = "Task" [ 1213.476646] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.484816] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a9a270a-8f4a-4471-b91e-c5a71cbc5e16 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.499490] env[61987]: DEBUG oslo_vmware.api [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062489, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.514246] env[61987]: DEBUG nova.compute.manager [req-9728c917-477a-40dd-bd79-34560324f44c req-d6100dc3-0dfc-4833-a3fa-1935a6eb965f service nova] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Detach interface failed, port_id=d3a24d05-4d6d-4ae9-9f18-33c775b215ee, reason: Instance b64111ab-4cd4-4584-b64e-e61b129a721d could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 1213.939973] env[61987]: INFO nova.compute.manager [-] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Took 1.43 seconds to deallocate network for instance. [ 1213.986797] env[61987]: DEBUG oslo_vmware.api [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062489, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.120616} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.987113] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1213.987312] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1213.987507] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1213.987692] env[61987]: INFO nova.compute.manager [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1213.987946] env[61987]: DEBUG oslo.service.loopingcall [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1213.988159] env[61987]: DEBUG nova.compute.manager [-] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1213.988253] env[61987]: DEBUG nova.network.neutron [-] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1214.143689] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0d450f58-52d0-4e2f-b196-e56813cfee6c tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "f030044b-bc6b-478c-8f63-d519aa76203c" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1214.144158] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0d450f58-52d0-4e2f-b196-e56813cfee6c tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "f030044b-bc6b-478c-8f63-d519aa76203c" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1214.485080] env[61987]: INFO nova.compute.manager [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Took 0.54 seconds to detach 1 volumes for instance. [ 1214.487311] env[61987]: DEBUG nova.compute.manager [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Deleting volume: 6e8a12d3-a1ae-481c-ad5f-a05364a8dc87 {{(pid=61987) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3281}} [ 1214.647959] env[61987]: INFO nova.compute.manager [None req-0d450f58-52d0-4e2f-b196-e56813cfee6c tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Detaching volume e6316fa6-4d06-40fb-90d8-60c4cfd7c454 [ 1214.680352] env[61987]: INFO nova.virt.block_device [None req-0d450f58-52d0-4e2f-b196-e56813cfee6c tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Attempting to driver detach volume e6316fa6-4d06-40fb-90d8-60c4cfd7c454 from mountpoint /dev/sdb [ 1214.680632] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d450f58-52d0-4e2f-b196-e56813cfee6c tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Volume detach. Driver type: vmdk {{(pid=61987) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1214.680852] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d450f58-52d0-4e2f-b196-e56813cfee6c tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234450', 'volume_id': 'e6316fa6-4d06-40fb-90d8-60c4cfd7c454', 'name': 'volume-e6316fa6-4d06-40fb-90d8-60c4cfd7c454', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'f030044b-bc6b-478c-8f63-d519aa76203c', 'attached_at': '', 'detached_at': '', 'volume_id': 'e6316fa6-4d06-40fb-90d8-60c4cfd7c454', 'serial': 'e6316fa6-4d06-40fb-90d8-60c4cfd7c454'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1214.681880] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8cfa137-5dd6-47ea-8cb5-3d76d3e24aba {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.703924] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43778b1a-45bd-428e-a32c-eacf2a135c90 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.710518] env[61987]: DEBUG nova.network.neutron [-] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1214.712431] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a3644b0-b79c-4c24-b97a-e96c605c6d12 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.733552] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0392fd1e-29b4-4f47-b96f-e5b46427c2c0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.748391] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d450f58-52d0-4e2f-b196-e56813cfee6c tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] The volume has not been displaced from its original location: [datastore1] volume-e6316fa6-4d06-40fb-90d8-60c4cfd7c454/volume-e6316fa6-4d06-40fb-90d8-60c4cfd7c454.vmdk. No consolidation needed. {{(pid=61987) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1214.753568] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d450f58-52d0-4e2f-b196-e56813cfee6c tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Reconfiguring VM instance instance-00000071 to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1214.754023] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aab3dd5f-3be6-4be9-af9d-c3bc7e1f4bd1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.771345] env[61987]: DEBUG oslo_vmware.api [None req-0d450f58-52d0-4e2f-b196-e56813cfee6c tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1214.771345] env[61987]: value = "task-1062491" [ 1214.771345] env[61987]: _type = "Task" [ 1214.771345] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.778698] env[61987]: DEBUG oslo_vmware.api [None req-0d450f58-52d0-4e2f-b196-e56813cfee6c tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062491, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.000112] env[61987]: DEBUG nova.compute.manager [req-834aee05-f37d-4c6e-a88c-e4115eef43c0 req-f69fe908-d435-41f5-a14b-c98aaa6c2007 service nova] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Received event network-vif-deleted-e48cc892-3cba-40a4-a803-6eab279a5d72 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1215.023966] env[61987]: DEBUG oslo_concurrency.lockutils [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1215.024227] env[61987]: DEBUG oslo_concurrency.lockutils [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1215.024442] env[61987]: DEBUG nova.objects.instance [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lazy-loading 'resources' on Instance uuid b64111ab-4cd4-4584-b64e-e61b129a721d {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1215.216677] env[61987]: INFO nova.compute.manager [-] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Took 1.23 seconds to deallocate network for instance. [ 1215.281177] env[61987]: DEBUG oslo_vmware.api [None req-0d450f58-52d0-4e2f-b196-e56813cfee6c tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062491, 'name': ReconfigVM_Task, 'duration_secs': 0.224654} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.281422] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d450f58-52d0-4e2f-b196-e56813cfee6c tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Reconfigured VM instance instance-00000071 to detach disk 2001 {{(pid=61987) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1215.285866] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93c9fe0e-1fd2-4ede-ba80-f73798eb9c70 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.300774] env[61987]: DEBUG oslo_vmware.api [None req-0d450f58-52d0-4e2f-b196-e56813cfee6c tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1215.300774] env[61987]: value = "task-1062492" [ 1215.300774] env[61987]: _type = "Task" [ 1215.300774] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.308021] env[61987]: DEBUG oslo_vmware.api [None req-0d450f58-52d0-4e2f-b196-e56813cfee6c tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062492, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.592185] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f3f3dd-d023-443c-b4e0-d65a106aa716 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.599825] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e93993a-571e-4e06-9817-2d02729e7a17 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.629101] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b84f9efb-0c32-47ca-95e2-35e0d32834df {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.636074] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c928cf3-c6ae-4bf4-a5bf-a9cc9e7d1e8d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.648882] env[61987]: DEBUG nova.compute.provider_tree [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1215.722300] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1215.811226] env[61987]: DEBUG oslo_vmware.api [None req-0d450f58-52d0-4e2f-b196-e56813cfee6c tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062492, 'name': ReconfigVM_Task, 'duration_secs': 0.12425} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.811501] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d450f58-52d0-4e2f-b196-e56813cfee6c tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-234450', 'volume_id': 'e6316fa6-4d06-40fb-90d8-60c4cfd7c454', 'name': 'volume-e6316fa6-4d06-40fb-90d8-60c4cfd7c454', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'f030044b-bc6b-478c-8f63-d519aa76203c', 'attached_at': '', 'detached_at': '', 'volume_id': 'e6316fa6-4d06-40fb-90d8-60c4cfd7c454', 'serial': 'e6316fa6-4d06-40fb-90d8-60c4cfd7c454'} {{(pid=61987) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1216.152222] env[61987]: DEBUG nova.scheduler.client.report [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1216.350221] env[61987]: DEBUG nova.objects.instance [None req-0d450f58-52d0-4e2f-b196-e56813cfee6c tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lazy-loading 'flavor' on Instance uuid f030044b-bc6b-478c-8f63-d519aa76203c {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1216.656781] env[61987]: DEBUG oslo_concurrency.lockutils [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.632s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1216.660041] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.937s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1216.660329] env[61987]: DEBUG nova.objects.instance [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lazy-loading 'resources' on Instance uuid 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1216.675647] env[61987]: INFO nova.scheduler.client.report [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Deleted allocations for instance b64111ab-4cd4-4584-b64e-e61b129a721d [ 1217.183057] env[61987]: DEBUG oslo_concurrency.lockutils [None req-588ebf70-42c2-4030-b8a3-065a1f0a550f tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "b64111ab-4cd4-4584-b64e-e61b129a721d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.413s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1217.183528] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2bd01824-9c8b-494a-b29d-a691c2f784a2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "b64111ab-4cd4-4584-b64e-e61b129a721d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 6.116s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1217.183754] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2bd01824-9c8b-494a-b29d-a691c2f784a2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "b64111ab-4cd4-4584-b64e-e61b129a721d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1217.183964] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2bd01824-9c8b-494a-b29d-a691c2f784a2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "b64111ab-4cd4-4584-b64e-e61b129a721d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1217.184160] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2bd01824-9c8b-494a-b29d-a691c2f784a2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "b64111ab-4cd4-4584-b64e-e61b129a721d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1217.185691] env[61987]: INFO nova.compute.manager [None req-2bd01824-9c8b-494a-b29d-a691c2f784a2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Terminating instance [ 1217.220480] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b8ae3b2-9283-4df9-99a9-789fbee884fc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.227954] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b190e0ef-791a-45a4-a8ff-d308b89cd8d7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.258547] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6e3882-74f9-4073-ad36-76387dcff8bd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.265612] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9adfa91-ace1-4be2-838c-420c4922fbf4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.278573] env[61987]: DEBUG nova.compute.provider_tree [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1217.356419] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0d450f58-52d0-4e2f-b196-e56813cfee6c tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "f030044b-bc6b-478c-8f63-d519aa76203c" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.212s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1217.691614] env[61987]: DEBUG nova.compute.manager [None req-2bd01824-9c8b-494a-b29d-a691c2f784a2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1217.692035] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f5ef2aae-ee74-4208-a24e-994e4bc1241b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.701187] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9601818-1731-498d-bdb8-5e93d435465e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.725015] env[61987]: WARNING nova.virt.vmwareapi.driver [None req-2bd01824-9c8b-494a-b29d-a691c2f784a2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance b64111ab-4cd4-4584-b64e-e61b129a721d could not be found. [ 1217.725215] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2bd01824-9c8b-494a-b29d-a691c2f784a2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1217.725481] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-67d244b7-8cef-4788-b098-c17966910bf4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.732540] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-136bb2e9-243f-46ae-a960-1b32bc87db3a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.754921] env[61987]: WARNING nova.virt.vmwareapi.vmops [None req-2bd01824-9c8b-494a-b29d-a691c2f784a2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b64111ab-4cd4-4584-b64e-e61b129a721d could not be found. [ 1217.755114] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2bd01824-9c8b-494a-b29d-a691c2f784a2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1217.755302] env[61987]: INFO nova.compute.manager [None req-2bd01824-9c8b-494a-b29d-a691c2f784a2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Took 0.06 seconds to destroy the instance on the hypervisor. [ 1217.755539] env[61987]: DEBUG oslo.service.loopingcall [None req-2bd01824-9c8b-494a-b29d-a691c2f784a2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1217.755753] env[61987]: DEBUG nova.compute.manager [-] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1217.755841] env[61987]: DEBUG nova.network.neutron [-] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1217.781441] env[61987]: DEBUG nova.scheduler.client.report [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1218.273061] env[61987]: DEBUG nova.network.neutron [-] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1218.285747] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.627s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.307233] env[61987]: INFO nova.scheduler.client.report [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Deleted allocations for instance 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3 [ 1218.378072] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "f030044b-bc6b-478c-8f63-d519aa76203c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1218.378072] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "f030044b-bc6b-478c-8f63-d519aa76203c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.378072] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "f030044b-bc6b-478c-8f63-d519aa76203c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1218.378072] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "f030044b-bc6b-478c-8f63-d519aa76203c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.378306] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "f030044b-bc6b-478c-8f63-d519aa76203c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.380352] env[61987]: INFO nova.compute.manager [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Terminating instance [ 1218.775292] env[61987]: INFO nova.compute.manager [-] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Took 1.02 seconds to deallocate network for instance. [ 1218.789602] env[61987]: WARNING nova.volume.cinder [None req-2bd01824-9c8b-494a-b29d-a691c2f784a2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Attachment 56f32164-6d6a-4930-aabc-0884079c3943 does not exist. Ignoring.: cinderclient.exceptions.NotFound: Volume attachment could not be found with filter: attachment_id = 56f32164-6d6a-4930-aabc-0884079c3943. (HTTP 404) (Request-ID: req-2bb35155-9420-46dc-8c34-329e0af9c78c) [ 1218.790068] env[61987]: INFO nova.compute.manager [None req-2bd01824-9c8b-494a-b29d-a691c2f784a2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Took 0.01 seconds to detach 1 volumes for instance. [ 1218.792871] env[61987]: DEBUG nova.compute.manager [None req-2bd01824-9c8b-494a-b29d-a691c2f784a2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Deleting volume: 6e8a12d3-a1ae-481c-ad5f-a05364a8dc87 {{(pid=61987) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3281}} [ 1218.808895] env[61987]: WARNING nova.compute.manager [None req-2bd01824-9c8b-494a-b29d-a691c2f784a2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Failed to delete volume: 6e8a12d3-a1ae-481c-ad5f-a05364a8dc87 due to Volume 6e8a12d3-a1ae-481c-ad5f-a05364a8dc87 could not be found.: nova.exception.VolumeNotFound: Volume 6e8a12d3-a1ae-481c-ad5f-a05364a8dc87 could not be found. [ 1218.816821] env[61987]: DEBUG oslo_concurrency.lockutils [None req-5e099c42-e34a-42de-b4e5-a5f9eb1b3a09 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.448s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.884423] env[61987]: DEBUG nova.compute.manager [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1218.884696] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1218.885705] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5d5617b-cb93-437d-a9e8-0a9de4aad317 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.894212] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1218.894598] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-14355361-ba2d-4107-aca5-912a8c4f2d36 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.900988] env[61987]: DEBUG oslo_vmware.api [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1218.900988] env[61987]: value = "task-1062493" [ 1218.900988] env[61987]: _type = "Task" [ 1218.900988] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.908940] env[61987]: DEBUG oslo_vmware.api [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062493, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.317652] env[61987]: INFO nova.compute.manager [None req-2bd01824-9c8b-494a-b29d-a691c2f784a2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Instance disappeared during terminate [ 1219.317913] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2bd01824-9c8b-494a-b29d-a691c2f784a2 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "b64111ab-4cd4-4584-b64e-e61b129a721d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.134s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1219.332692] env[61987]: DEBUG oslo_concurrency.lockutils [None req-669831c6-4345-4482-b3c6-61a166a88fbe tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "0704efaf-08de-4888-a092-476bcb789eea" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1219.332935] env[61987]: DEBUG oslo_concurrency.lockutils [None req-669831c6-4345-4482-b3c6-61a166a88fbe tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "0704efaf-08de-4888-a092-476bcb789eea" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1219.333142] env[61987]: DEBUG nova.compute.manager [None req-669831c6-4345-4482-b3c6-61a166a88fbe tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1219.334228] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d05b4614-3602-49ab-bf8e-411fb688e09b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.342029] env[61987]: DEBUG nova.compute.manager [None req-669831c6-4345-4482-b3c6-61a166a88fbe tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61987) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1219.342029] env[61987]: DEBUG nova.objects.instance [None req-669831c6-4345-4482-b3c6-61a166a88fbe tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lazy-loading 'flavor' on Instance uuid 0704efaf-08de-4888-a092-476bcb789eea {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1219.411228] env[61987]: DEBUG oslo_vmware.api [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062493, 'name': PowerOffVM_Task, 'duration_secs': 0.204618} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.411492] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1219.411668] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1219.411920] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4b9ba4b8-ba4c-42b4-81c9-aea3aa61fff1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.476899] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1219.477277] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1219.477549] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Deleting the datastore file [datastore2] f030044b-bc6b-478c-8f63-d519aa76203c {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1219.477958] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5f20bdf5-bf88-4621-8b3b-41c49d22fcc9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.485555] env[61987]: DEBUG oslo_vmware.api [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1219.485555] env[61987]: value = "task-1062496" [ 1219.485555] env[61987]: _type = "Task" [ 1219.485555] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.493877] env[61987]: DEBUG oslo_vmware.api [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062496, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.995225] env[61987]: DEBUG oslo_vmware.api [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062496, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.121094} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.995510] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1219.995686] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1219.995865] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1219.996060] env[61987]: INFO nova.compute.manager [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1219.996311] env[61987]: DEBUG oslo.service.loopingcall [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1219.996513] env[61987]: DEBUG nova.compute.manager [-] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1219.996612] env[61987]: DEBUG nova.network.neutron [-] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1220.348612] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-669831c6-4345-4482-b3c6-61a166a88fbe tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1220.348962] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e2794af-2307-4136-abad-2e35a7779e5d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.356099] env[61987]: DEBUG oslo_vmware.api [None req-669831c6-4345-4482-b3c6-61a166a88fbe tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1220.356099] env[61987]: value = "task-1062497" [ 1220.356099] env[61987]: _type = "Task" [ 1220.356099] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.367146] env[61987]: DEBUG oslo_vmware.api [None req-669831c6-4345-4482-b3c6-61a166a88fbe tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062497, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.514334] env[61987]: DEBUG nova.compute.manager [req-7ecceeb7-1032-41f6-8832-c1ae63dc0b27 req-1b62b08a-6b25-43ac-8904-1c3b460bdc45 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Received event network-vif-deleted-2d38bac7-bebb-462a-aac9-bbff9d711577 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1220.514553] env[61987]: INFO nova.compute.manager [req-7ecceeb7-1032-41f6-8832-c1ae63dc0b27 req-1b62b08a-6b25-43ac-8904-1c3b460bdc45 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Neutron deleted interface 2d38bac7-bebb-462a-aac9-bbff9d711577; detaching it from the instance and deleting it from the info cache [ 1220.514776] env[61987]: DEBUG nova.network.neutron [req-7ecceeb7-1032-41f6-8832-c1ae63dc0b27 req-1b62b08a-6b25-43ac-8904-1c3b460bdc45 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1220.865392] env[61987]: DEBUG oslo_vmware.api [None req-669831c6-4345-4482-b3c6-61a166a88fbe tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062497, 'name': PowerOffVM_Task, 'duration_secs': 0.237827} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.865653] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-669831c6-4345-4482-b3c6-61a166a88fbe tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1220.865806] env[61987]: DEBUG nova.compute.manager [None req-669831c6-4345-4482-b3c6-61a166a88fbe tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1220.866939] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcd5449f-cfe8-4653-9b79-ae3f8a57a386 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.994202] env[61987]: DEBUG nova.network.neutron [-] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1221.016893] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3776ba5d-6c0e-46ce-80ac-adac828984fb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.026313] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a35a9ae-f26a-4932-8732-ca03f3e2d6f6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.049830] env[61987]: DEBUG nova.compute.manager [req-7ecceeb7-1032-41f6-8832-c1ae63dc0b27 req-1b62b08a-6b25-43ac-8904-1c3b460bdc45 service nova] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Detach interface failed, port_id=2d38bac7-bebb-462a-aac9-bbff9d711577, reason: Instance f030044b-bc6b-478c-8f63-d519aa76203c could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 1221.379764] env[61987]: DEBUG oslo_concurrency.lockutils [None req-669831c6-4345-4482-b3c6-61a166a88fbe tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "0704efaf-08de-4888-a092-476bcb789eea" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.046s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1221.497082] env[61987]: INFO nova.compute.manager [-] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Took 1.50 seconds to deallocate network for instance. [ 1221.776611] env[61987]: DEBUG nova.objects.instance [None req-79f18aee-2253-4f31-b505-f8d3f5506d80 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lazy-loading 'flavor' on Instance uuid 0704efaf-08de-4888-a092-476bcb789eea {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1222.005124] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1222.005124] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1222.005389] env[61987]: DEBUG nova.objects.instance [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lazy-loading 'resources' on Instance uuid f030044b-bc6b-478c-8f63-d519aa76203c {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1222.066933] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "4d15848d-5265-4897-bfa7-f3ad4c02ed83" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1222.067317] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "4d15848d-5265-4897-bfa7-f3ad4c02ed83" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1222.067382] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "4d15848d-5265-4897-bfa7-f3ad4c02ed83-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1222.067567] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "4d15848d-5265-4897-bfa7-f3ad4c02ed83-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1222.067749] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "4d15848d-5265-4897-bfa7-f3ad4c02ed83-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1222.069751] env[61987]: INFO nova.compute.manager [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Terminating instance [ 1222.282938] env[61987]: DEBUG oslo_concurrency.lockutils [None req-79f18aee-2253-4f31-b505-f8d3f5506d80 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "refresh_cache-0704efaf-08de-4888-a092-476bcb789eea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1222.283103] env[61987]: DEBUG oslo_concurrency.lockutils [None req-79f18aee-2253-4f31-b505-f8d3f5506d80 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "refresh_cache-0704efaf-08de-4888-a092-476bcb789eea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1222.283280] env[61987]: DEBUG nova.network.neutron [None req-79f18aee-2253-4f31-b505-f8d3f5506d80 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1222.283504] env[61987]: DEBUG nova.objects.instance [None req-79f18aee-2253-4f31-b505-f8d3f5506d80 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lazy-loading 'info_cache' on Instance uuid 0704efaf-08de-4888-a092-476bcb789eea {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1222.559736] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bbb4115-e625-4833-bc5a-a10cedcdb580 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.568673] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e94d6b49-0d65-4a64-ac46-848ec0dfb0c4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.573346] env[61987]: DEBUG nova.compute.manager [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1222.573609] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1222.574697] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7823708b-d925-4b0c-8d63-3167eeee2bbe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.600603] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d21113-e56b-4d7e-8b96-d7a75973378f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.604412] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1222.604889] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-af827e7a-574b-490d-a6ea-7619c84bb55f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.608986] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99176022-54b3-42c5-8e05-67066340f145 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.613493] env[61987]: DEBUG oslo_vmware.api [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1222.613493] env[61987]: value = "task-1062498" [ 1222.613493] env[61987]: _type = "Task" [ 1222.613493] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.623831] env[61987]: DEBUG nova.compute.provider_tree [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1222.629729] env[61987]: DEBUG oslo_vmware.api [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062498, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.786972] env[61987]: DEBUG nova.objects.base [None req-79f18aee-2253-4f31-b505-f8d3f5506d80 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Object Instance<0704efaf-08de-4888-a092-476bcb789eea> lazy-loaded attributes: flavor,info_cache {{(pid=61987) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1223.124174] env[61987]: DEBUG oslo_vmware.api [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062498, 'name': PowerOffVM_Task, 'duration_secs': 0.208919} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.124546] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1223.124595] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1223.124809] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3fd7f643-eb5a-4cd7-a24e-712dbc620c07 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.142538] env[61987]: ERROR nova.scheduler.client.report [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [req-6fcadcd4-f3c2-4ba4-a9cf-348b1a3bd64f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8db744e5-1e14-4e98-9818-695ea4b7dfd0. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-6fcadcd4-f3c2-4ba4-a9cf-348b1a3bd64f"}]} [ 1223.158155] env[61987]: DEBUG nova.scheduler.client.report [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Refreshing inventories for resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1223.171030] env[61987]: DEBUG nova.scheduler.client.report [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Updating ProviderTree inventory for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1223.171313] env[61987]: DEBUG nova.compute.provider_tree [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1223.183692] env[61987]: DEBUG nova.scheduler.client.report [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Refreshing aggregate associations for resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0, aggregates: None {{(pid=61987) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1223.188573] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1223.188811] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1223.189019] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Deleting the datastore file [datastore2] 4d15848d-5265-4897-bfa7-f3ad4c02ed83 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1223.189292] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-358aa036-9e41-4cc7-88d6-ee5d8120f190 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.194830] env[61987]: DEBUG oslo_vmware.api [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for the task: (returnval){ [ 1223.194830] env[61987]: value = "task-1062500" [ 1223.194830] env[61987]: _type = "Task" [ 1223.194830] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.202728] env[61987]: DEBUG oslo_vmware.api [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062500, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.203604] env[61987]: DEBUG nova.scheduler.client.report [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Refreshing trait associations for resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61987) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1223.248703] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be9cf15-b77a-4e64-895b-e6e6f680bff7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.255763] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca14b1f-9a29-40ab-9186-960916ee0725 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.285703] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0db01a41-ac54-4da0-9cd2-8ca27b0a89f7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.302865] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cdab93c-3c74-440f-bed0-81ec6f7b093e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.316280] env[61987]: DEBUG nova.compute.provider_tree [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1223.501834] env[61987]: DEBUG nova.network.neutron [None req-79f18aee-2253-4f31-b505-f8d3f5506d80 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Updating instance_info_cache with network_info: [{"id": "79e1d3b3-41db-4650-95bd-11df0dc06e80", "address": "fa:16:3e:7f:09:fd", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79e1d3b3-41", "ovs_interfaceid": "79e1d3b3-41db-4650-95bd-11df0dc06e80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1223.705307] env[61987]: DEBUG oslo_vmware.api [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Task: {'id': task-1062500, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156714} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.705547] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1223.705715] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1223.705898] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1223.706095] env[61987]: INFO nova.compute.manager [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1223.706349] env[61987]: DEBUG oslo.service.loopingcall [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1223.706549] env[61987]: DEBUG nova.compute.manager [-] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1223.706669] env[61987]: DEBUG nova.network.neutron [-] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1223.850049] env[61987]: DEBUG nova.scheduler.client.report [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Updated inventory for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with generation 151 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1223.850385] env[61987]: DEBUG nova.compute.provider_tree [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Updating resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 generation from 151 to 152 during operation: update_inventory {{(pid=61987) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1223.850589] env[61987]: DEBUG nova.compute.provider_tree [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1223.978241] env[61987]: DEBUG nova.compute.manager [req-6db315e1-0da6-46d2-9835-64ae2e846c8b req-291c2d15-bfac-4c08-95db-54fa83c9ac1b service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Received event network-vif-deleted-203d81e4-dbb8-4685-a463-9de48b835ff7 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1223.978461] env[61987]: INFO nova.compute.manager [req-6db315e1-0da6-46d2-9835-64ae2e846c8b req-291c2d15-bfac-4c08-95db-54fa83c9ac1b service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Neutron deleted interface 203d81e4-dbb8-4685-a463-9de48b835ff7; detaching it from the instance and deleting it from the info cache [ 1223.978640] env[61987]: DEBUG nova.network.neutron [req-6db315e1-0da6-46d2-9835-64ae2e846c8b req-291c2d15-bfac-4c08-95db-54fa83c9ac1b service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1224.007950] env[61987]: DEBUG oslo_concurrency.lockutils [None req-79f18aee-2253-4f31-b505-f8d3f5506d80 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "refresh_cache-0704efaf-08de-4888-a092-476bcb789eea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1224.355738] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.351s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1224.375552] env[61987]: INFO nova.scheduler.client.report [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Deleted allocations for instance f030044b-bc6b-478c-8f63-d519aa76203c [ 1224.455838] env[61987]: DEBUG nova.network.neutron [-] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1224.481541] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f274beb8-f5f4-4dbd-8666-173929639776 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.490604] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd080c2b-1c52-4040-867c-3d8c76f104e8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.515802] env[61987]: DEBUG nova.compute.manager [req-6db315e1-0da6-46d2-9835-64ae2e846c8b req-291c2d15-bfac-4c08-95db-54fa83c9ac1b service nova] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Detach interface failed, port_id=203d81e4-dbb8-4685-a463-9de48b835ff7, reason: Instance 4d15848d-5265-4897-bfa7-f3ad4c02ed83 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 1224.885432] env[61987]: DEBUG oslo_concurrency.lockutils [None req-64d9c6b6-f16d-4f47-a464-9a2f92fb36ff tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "f030044b-bc6b-478c-8f63-d519aa76203c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.508s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1224.958451] env[61987]: INFO nova.compute.manager [-] [instance: 4d15848d-5265-4897-bfa7-f3ad4c02ed83] Took 1.25 seconds to deallocate network for instance. [ 1225.017856] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-79f18aee-2253-4f31-b505-f8d3f5506d80 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1225.018197] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-30b7bb54-44e8-46b2-a1c8-86f0751797d8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.026154] env[61987]: DEBUG oslo_vmware.api [None req-79f18aee-2253-4f31-b505-f8d3f5506d80 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1225.026154] env[61987]: value = "task-1062501" [ 1225.026154] env[61987]: _type = "Task" [ 1225.026154] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.033942] env[61987]: DEBUG oslo_vmware.api [None req-79f18aee-2253-4f31-b505-f8d3f5506d80 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062501, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.464835] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1225.464835] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1225.465213] env[61987]: DEBUG nova.objects.instance [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lazy-loading 'resources' on Instance uuid 4d15848d-5265-4897-bfa7-f3ad4c02ed83 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1225.536115] env[61987]: DEBUG oslo_vmware.api [None req-79f18aee-2253-4f31-b505-f8d3f5506d80 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062501, 'name': PowerOnVM_Task, 'duration_secs': 0.41875} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.536401] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-79f18aee-2253-4f31-b505-f8d3f5506d80 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1225.536607] env[61987]: DEBUG nova.compute.manager [None req-79f18aee-2253-4f31-b505-f8d3f5506d80 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1225.537418] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf4a1095-0817-466b-8676-da2abd906ce7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.010775] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc800934-9313-46eb-9987-b1d5a9561a71 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.018348] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb98f82c-a989-4215-bb1b-3b184ca13833 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.051799] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-170dd040-2751-415e-b2c0-12f11ba2d112 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.060605] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-111c265c-ed8e-4d67-a894-eb36f7fadef4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.075251] env[61987]: DEBUG nova.compute.provider_tree [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1226.578206] env[61987]: DEBUG nova.scheduler.client.report [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1226.616088] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "054840f7-a1f2-4888-9a04-75c788dee325" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1226.616317] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "054840f7-a1f2-4888-9a04-75c788dee325" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1226.859830] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ec47303-c005-47df-b476-5594ba21bd8c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.866429] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7812e2a0-97b1-40ee-8268-f7edc9b9ba55 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Suspending the VM {{(pid=61987) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1226.866664] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-6247319a-25e1-4f0d-afd1-13fc1ee59dc7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.873238] env[61987]: DEBUG oslo_vmware.api [None req-7812e2a0-97b1-40ee-8268-f7edc9b9ba55 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1226.873238] env[61987]: value = "task-1062502" [ 1226.873238] env[61987]: _type = "Task" [ 1226.873238] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.880561] env[61987]: DEBUG oslo_vmware.api [None req-7812e2a0-97b1-40ee-8268-f7edc9b9ba55 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062502, 'name': SuspendVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.083261] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.618s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1227.100456] env[61987]: INFO nova.scheduler.client.report [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Deleted allocations for instance 4d15848d-5265-4897-bfa7-f3ad4c02ed83 [ 1227.119050] env[61987]: DEBUG nova.compute.manager [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1227.383260] env[61987]: DEBUG oslo_vmware.api [None req-7812e2a0-97b1-40ee-8268-f7edc9b9ba55 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062502, 'name': SuspendVM_Task} progress is 100%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.609077] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a5fccc24-1ee9-4a2d-b4a2-98b0c80b02a4 tempest-ServerActionsTestOtherA-1972265026 tempest-ServerActionsTestOtherA-1972265026-project-member] Lock "4d15848d-5265-4897-bfa7-f3ad4c02ed83" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.542s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1227.636896] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1227.637171] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1227.640187] env[61987]: INFO nova.compute.claims [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1227.883196] env[61987]: DEBUG oslo_vmware.api [None req-7812e2a0-97b1-40ee-8268-f7edc9b9ba55 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062502, 'name': SuspendVM_Task, 'duration_secs': 0.546008} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.883468] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-7812e2a0-97b1-40ee-8268-f7edc9b9ba55 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Suspended the VM {{(pid=61987) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1227.883654] env[61987]: DEBUG nova.compute.manager [None req-7812e2a0-97b1-40ee-8268-f7edc9b9ba55 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1227.884425] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda7dd8b-439b-4cad-a0ff-dcd3aa8c9083 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.694891] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ddb190c-8879-49b5-bc0e-4ab30e8cfc8e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.703210] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adaecea1-32fd-4169-95cf-489bec322bcf {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.736724] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6acfce62-c7fc-4829-aaac-6541a4edf946 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.744194] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d68b998d-3c7f-49a0-9c98-15fdb90306f6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.760715] env[61987]: DEBUG nova.compute.provider_tree [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1229.263085] env[61987]: INFO nova.compute.manager [None req-b6ba6bb8-cc19-4b66-ba4a-f50afec6e63e tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Resuming [ 1229.263744] env[61987]: DEBUG nova.objects.instance [None req-b6ba6bb8-cc19-4b66-ba4a-f50afec6e63e tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lazy-loading 'flavor' on Instance uuid 0704efaf-08de-4888-a092-476bcb789eea {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1229.265625] env[61987]: DEBUG nova.scheduler.client.report [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1229.771588] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.134s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.772111] env[61987]: DEBUG nova.compute.manager [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1230.277704] env[61987]: DEBUG nova.compute.utils [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1230.279241] env[61987]: DEBUG nova.compute.manager [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Allocating IP information in the background. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1230.279888] env[61987]: DEBUG nova.network.neutron [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] allocate_for_instance() {{(pid=61987) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1230.336123] env[61987]: DEBUG nova.policy [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eef90ff5a24b4091ac65ceef920778c1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '90faa728219a432fa241b635071729e5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61987) authorize /opt/stack/nova/nova/policy.py:201}} [ 1230.596521] env[61987]: DEBUG nova.network.neutron [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Successfully created port: 7906310e-5bc1-4ad3-85fb-062c271a2f85 {{(pid=61987) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1230.783415] env[61987]: DEBUG nova.compute.manager [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1230.786458] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b6ba6bb8-cc19-4b66-ba4a-f50afec6e63e tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "refresh_cache-0704efaf-08de-4888-a092-476bcb789eea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1230.786626] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b6ba6bb8-cc19-4b66-ba4a-f50afec6e63e tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquired lock "refresh_cache-0704efaf-08de-4888-a092-476bcb789eea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1230.786801] env[61987]: DEBUG nova.network.neutron [None req-b6ba6bb8-cc19-4b66-ba4a-f50afec6e63e tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1231.183025] env[61987]: DEBUG oslo_concurrency.lockutils [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquiring lock "9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1231.183025] env[61987]: DEBUG oslo_concurrency.lockutils [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1231.562191] env[61987]: DEBUG nova.network.neutron [None req-b6ba6bb8-cc19-4b66-ba4a-f50afec6e63e tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Updating instance_info_cache with network_info: [{"id": "79e1d3b3-41db-4650-95bd-11df0dc06e80", "address": "fa:16:3e:7f:09:fd", "network": {"id": "408db618-a82c-4707-8e98-c95955ea0723", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1825523578-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.140", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "764082416d314c3f92eb83f576aef222", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a111da1-447a-4722-9662-b0d6a2886a65", "external-id": "nsx-vlan-transportzone-435", "segmentation_id": 435, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79e1d3b3-41", "ovs_interfaceid": "79e1d3b3-41db-4650-95bd-11df0dc06e80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1231.571121] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquiring lock "03766f52-c102-40ff-b004-7748231d0a03" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1231.571456] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "03766f52-c102-40ff-b004-7748231d0a03" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1231.685236] env[61987]: DEBUG nova.compute.manager [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1231.797267] env[61987]: DEBUG nova.compute.manager [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1231.822680] env[61987]: DEBUG nova.virt.hardware [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1231.822948] env[61987]: DEBUG nova.virt.hardware [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1231.823132] env[61987]: DEBUG nova.virt.hardware [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1231.823343] env[61987]: DEBUG nova.virt.hardware [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1231.823542] env[61987]: DEBUG nova.virt.hardware [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1231.823774] env[61987]: DEBUG nova.virt.hardware [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1231.824029] env[61987]: DEBUG nova.virt.hardware [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1231.824213] env[61987]: DEBUG nova.virt.hardware [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1231.824393] env[61987]: DEBUG nova.virt.hardware [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1231.824561] env[61987]: DEBUG nova.virt.hardware [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1231.824742] env[61987]: DEBUG nova.virt.hardware [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1231.825635] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-558bc6ee-f8bd-4d3a-84de-fb17f056c100 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.833871] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e67a37c-d851-4d93-a4c7-f1e143f8daba {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.066086] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b6ba6bb8-cc19-4b66-ba4a-f50afec6e63e tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Releasing lock "refresh_cache-0704efaf-08de-4888-a092-476bcb789eea" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1232.067010] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d050502-5ce1-4ce8-a21c-71af748dd620 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.073810] env[61987]: DEBUG nova.compute.manager [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Starting instance... {{(pid=61987) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1232.076391] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b6ba6bb8-cc19-4b66-ba4a-f50afec6e63e tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Resuming the VM {{(pid=61987) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1184}} [ 1232.076817] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-660e2ef8-1e21-4d38-801c-f39597a5cf96 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.085076] env[61987]: DEBUG oslo_vmware.api [None req-b6ba6bb8-cc19-4b66-ba4a-f50afec6e63e tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1232.085076] env[61987]: value = "task-1062503" [ 1232.085076] env[61987]: _type = "Task" [ 1232.085076] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.092921] env[61987]: DEBUG oslo_vmware.api [None req-b6ba6bb8-cc19-4b66-ba4a-f50afec6e63e tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062503, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.207098] env[61987]: DEBUG oslo_concurrency.lockutils [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1232.207382] env[61987]: DEBUG oslo_concurrency.lockutils [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1232.209102] env[61987]: INFO nova.compute.claims [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1232.264042] env[61987]: DEBUG nova.compute.manager [req-20ca0424-809f-4e15-9e79-3f86fe80c792 req-032f1b25-d03d-4d25-9d3d-ac9281795c5c service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Received event network-vif-plugged-7906310e-5bc1-4ad3-85fb-062c271a2f85 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1232.264265] env[61987]: DEBUG oslo_concurrency.lockutils [req-20ca0424-809f-4e15-9e79-3f86fe80c792 req-032f1b25-d03d-4d25-9d3d-ac9281795c5c service nova] Acquiring lock "054840f7-a1f2-4888-9a04-75c788dee325-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1232.264487] env[61987]: DEBUG oslo_concurrency.lockutils [req-20ca0424-809f-4e15-9e79-3f86fe80c792 req-032f1b25-d03d-4d25-9d3d-ac9281795c5c service nova] Lock "054840f7-a1f2-4888-9a04-75c788dee325-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1232.264667] env[61987]: DEBUG oslo_concurrency.lockutils [req-20ca0424-809f-4e15-9e79-3f86fe80c792 req-032f1b25-d03d-4d25-9d3d-ac9281795c5c service nova] Lock "054840f7-a1f2-4888-9a04-75c788dee325-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1232.264843] env[61987]: DEBUG nova.compute.manager [req-20ca0424-809f-4e15-9e79-3f86fe80c792 req-032f1b25-d03d-4d25-9d3d-ac9281795c5c service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] No waiting events found dispatching network-vif-plugged-7906310e-5bc1-4ad3-85fb-062c271a2f85 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1232.265134] env[61987]: WARNING nova.compute.manager [req-20ca0424-809f-4e15-9e79-3f86fe80c792 req-032f1b25-d03d-4d25-9d3d-ac9281795c5c service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Received unexpected event network-vif-plugged-7906310e-5bc1-4ad3-85fb-062c271a2f85 for instance with vm_state building and task_state spawning. [ 1232.346983] env[61987]: DEBUG nova.network.neutron [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Successfully updated port: 7906310e-5bc1-4ad3-85fb-062c271a2f85 {{(pid=61987) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1232.594685] env[61987]: DEBUG oslo_vmware.api [None req-b6ba6bb8-cc19-4b66-ba4a-f50afec6e63e tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062503, 'name': PowerOnVM_Task, 'duration_secs': 0.493724} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.594984] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b6ba6bb8-cc19-4b66-ba4a-f50afec6e63e tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Resumed the VM {{(pid=61987) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1189}} [ 1232.595507] env[61987]: DEBUG nova.compute.manager [None req-b6ba6bb8-cc19-4b66-ba4a-f50afec6e63e tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1232.596692] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1232.597554] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b30d78b-9d83-41f7-a5c9-752ccb85390a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.850393] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1232.850773] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquired lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1232.850773] env[61987]: DEBUG nova.network.neutron [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1233.271139] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06e8f0fe-ea31-485a-b5aa-62ecf07b6f92 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.279142] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6eaa816-39f2-432e-bfe5-61736c8aa422 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.309643] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b23c6e96-1755-4b6e-81a2-4ef72d1e4180 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.316957] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94077a36-e01e-4fdd-975e-b87af9f641fd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.330023] env[61987]: DEBUG nova.compute.provider_tree [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1233.394131] env[61987]: DEBUG nova.network.neutron [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1233.528757] env[61987]: DEBUG nova.network.neutron [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Updating instance_info_cache with network_info: [{"id": "7906310e-5bc1-4ad3-85fb-062c271a2f85", "address": "fa:16:3e:99:f5:8a", "network": {"id": "eac55241-aad5-4ddb-a01e-625c9298a8af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1411727225-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90faa728219a432fa241b635071729e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f996252-e329-42bd-a897-446dfe2b81cd", "external-id": "nsx-vlan-transportzone-535", "segmentation_id": 535, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7906310e-5b", "ovs_interfaceid": "7906310e-5bc1-4ad3-85fb-062c271a2f85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1233.833521] env[61987]: DEBUG nova.scheduler.client.report [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1234.031370] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Releasing lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1234.031748] env[61987]: DEBUG nova.compute.manager [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Instance network_info: |[{"id": "7906310e-5bc1-4ad3-85fb-062c271a2f85", "address": "fa:16:3e:99:f5:8a", "network": {"id": "eac55241-aad5-4ddb-a01e-625c9298a8af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1411727225-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90faa728219a432fa241b635071729e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f996252-e329-42bd-a897-446dfe2b81cd", "external-id": "nsx-vlan-transportzone-535", "segmentation_id": 535, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7906310e-5b", "ovs_interfaceid": "7906310e-5bc1-4ad3-85fb-062c271a2f85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1234.032141] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:f5:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1f996252-e329-42bd-a897-446dfe2b81cd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7906310e-5bc1-4ad3-85fb-062c271a2f85', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1234.041307] env[61987]: DEBUG oslo.service.loopingcall [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1234.041527] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1234.041750] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d5857af3-dec2-485f-9c01-c8b38da53c74 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.061545] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1234.061545] env[61987]: value = "task-1062504" [ 1234.061545] env[61987]: _type = "Task" [ 1234.061545] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.068716] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062504, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.083210] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "0704efaf-08de-4888-a092-476bcb789eea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1234.083438] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "0704efaf-08de-4888-a092-476bcb789eea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.083644] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "0704efaf-08de-4888-a092-476bcb789eea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1234.083832] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "0704efaf-08de-4888-a092-476bcb789eea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.084021] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "0704efaf-08de-4888-a092-476bcb789eea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1234.085996] env[61987]: INFO nova.compute.manager [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Terminating instance [ 1234.287792] env[61987]: DEBUG nova.compute.manager [req-d9c652ca-025b-47ce-b3a6-35c47bc87f50 req-3f31a548-ea85-4de4-8229-356eca9b838e service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Received event network-changed-7906310e-5bc1-4ad3-85fb-062c271a2f85 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1234.287992] env[61987]: DEBUG nova.compute.manager [req-d9c652ca-025b-47ce-b3a6-35c47bc87f50 req-3f31a548-ea85-4de4-8229-356eca9b838e service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Refreshing instance network info cache due to event network-changed-7906310e-5bc1-4ad3-85fb-062c271a2f85. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1234.288213] env[61987]: DEBUG oslo_concurrency.lockutils [req-d9c652ca-025b-47ce-b3a6-35c47bc87f50 req-3f31a548-ea85-4de4-8229-356eca9b838e service nova] Acquiring lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1234.288373] env[61987]: DEBUG oslo_concurrency.lockutils [req-d9c652ca-025b-47ce-b3a6-35c47bc87f50 req-3f31a548-ea85-4de4-8229-356eca9b838e service nova] Acquired lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1234.288550] env[61987]: DEBUG nova.network.neutron [req-d9c652ca-025b-47ce-b3a6-35c47bc87f50 req-3f31a548-ea85-4de4-8229-356eca9b838e service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Refreshing network info cache for port 7906310e-5bc1-4ad3-85fb-062c271a2f85 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1234.338849] env[61987]: DEBUG oslo_concurrency.lockutils [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.131s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1234.339498] env[61987]: DEBUG nova.compute.manager [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1234.342881] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.746s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.343769] env[61987]: INFO nova.compute.claims [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1234.572953] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062504, 'name': CreateVM_Task, 'duration_secs': 0.314914} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.573139] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1234.573850] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1234.574041] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1234.574434] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1234.574699] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb880467-f481-482b-a2c8-8700f1b33d6b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.579053] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1234.579053] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5221f29d-b2f5-0da5-4321-e76f0d21d1c9" [ 1234.579053] env[61987]: _type = "Task" [ 1234.579053] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.586196] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5221f29d-b2f5-0da5-4321-e76f0d21d1c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.588840] env[61987]: DEBUG nova.compute.manager [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1234.589103] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1234.589811] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aec4b08-3ece-492f-9948-92f94fdd4897 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.595825] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1234.596053] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e1a81bf7-ea72-49d2-9f78-4dcf1bedd1c5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.601823] env[61987]: DEBUG oslo_vmware.api [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1234.601823] env[61987]: value = "task-1062505" [ 1234.601823] env[61987]: _type = "Task" [ 1234.601823] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.609307] env[61987]: DEBUG oslo_vmware.api [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062505, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.848183] env[61987]: DEBUG nova.compute.utils [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1234.851680] env[61987]: DEBUG nova.compute.manager [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Not allocating networking since 'none' was specified. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 1235.024688] env[61987]: DEBUG nova.network.neutron [req-d9c652ca-025b-47ce-b3a6-35c47bc87f50 req-3f31a548-ea85-4de4-8229-356eca9b838e service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Updated VIF entry in instance network info cache for port 7906310e-5bc1-4ad3-85fb-062c271a2f85. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1235.025215] env[61987]: DEBUG nova.network.neutron [req-d9c652ca-025b-47ce-b3a6-35c47bc87f50 req-3f31a548-ea85-4de4-8229-356eca9b838e service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Updating instance_info_cache with network_info: [{"id": "7906310e-5bc1-4ad3-85fb-062c271a2f85", "address": "fa:16:3e:99:f5:8a", "network": {"id": "eac55241-aad5-4ddb-a01e-625c9298a8af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1411727225-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90faa728219a432fa241b635071729e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f996252-e329-42bd-a897-446dfe2b81cd", "external-id": "nsx-vlan-transportzone-535", "segmentation_id": 535, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7906310e-5b", "ovs_interfaceid": "7906310e-5bc1-4ad3-85fb-062c271a2f85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1235.089436] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5221f29d-b2f5-0da5-4321-e76f0d21d1c9, 'name': SearchDatastore_Task, 'duration_secs': 0.009327} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.089876] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1235.089948] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1235.090146] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1235.090299] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1235.090482] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1235.090739] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ffba3a10-3ab2-41ef-bbb5-851d5f6450c4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.098731] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1235.098940] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1235.099842] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f4b4ebd-37f9-479e-a5a7-cea3cf18e2dd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.107368] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1235.107368] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52c7960b-2e58-9de2-d8c9-7cfbf6d97580" [ 1235.107368] env[61987]: _type = "Task" [ 1235.107368] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.110373] env[61987]: DEBUG oslo_vmware.api [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062505, 'name': PowerOffVM_Task, 'duration_secs': 0.189279} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.114579] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1235.114775] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1235.115012] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-049dd1d1-0b60-4c7a-b1bb-9417946093e3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.121226] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c7960b-2e58-9de2-d8c9-7cfbf6d97580, 'name': SearchDatastore_Task, 'duration_secs': 0.008506} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.121906] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-304fc7ff-2ba6-4947-babb-b44b2970ebb3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.126529] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1235.126529] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52e244fb-f480-f785-1e43-aa9980130b00" [ 1235.126529] env[61987]: _type = "Task" [ 1235.126529] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.134389] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52e244fb-f480-f785-1e43-aa9980130b00, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.180858] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1235.181094] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Deleting contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1235.181297] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Deleting the datastore file [datastore1] 0704efaf-08de-4888-a092-476bcb789eea {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1235.181565] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5e94e447-a78c-4aab-8b95-05e581807f87 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.187538] env[61987]: DEBUG oslo_vmware.api [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for the task: (returnval){ [ 1235.187538] env[61987]: value = "task-1062507" [ 1235.187538] env[61987]: _type = "Task" [ 1235.187538] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.195385] env[61987]: DEBUG oslo_vmware.api [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062507, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.352429] env[61987]: DEBUG nova.compute.manager [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1235.415335] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04c66f05-82d1-430c-be11-779fed9acd9d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.423319] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-761da30b-4b1f-4e55-bb66-8fd68c8e30bc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.452255] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c16d1d-fb9b-4c40-8743-c52137cf96e3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.459087] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f500ac3f-22e5-4024-b3ef-8818dca10492 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.471826] env[61987]: DEBUG nova.compute.provider_tree [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1235.528293] env[61987]: DEBUG oslo_concurrency.lockutils [req-d9c652ca-025b-47ce-b3a6-35c47bc87f50 req-3f31a548-ea85-4de4-8229-356eca9b838e service nova] Releasing lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1235.637742] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52e244fb-f480-f785-1e43-aa9980130b00, 'name': SearchDatastore_Task, 'duration_secs': 0.00837} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.637958] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1235.638239] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 054840f7-a1f2-4888-9a04-75c788dee325/054840f7-a1f2-4888-9a04-75c788dee325.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1235.638507] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c73e2f6d-6ddc-4421-850b-a087daffd052 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.645306] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1235.645306] env[61987]: value = "task-1062508" [ 1235.645306] env[61987]: _type = "Task" [ 1235.645306] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.654104] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062508, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.697022] env[61987]: DEBUG oslo_vmware.api [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Task: {'id': task-1062507, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158831} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.697311] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1235.697523] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Deleted contents of the VM from datastore datastore1 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1235.697690] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1235.697880] env[61987]: INFO nova.compute.manager [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1235.698152] env[61987]: DEBUG oslo.service.loopingcall [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1235.698369] env[61987]: DEBUG nova.compute.manager [-] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1235.698462] env[61987]: DEBUG nova.network.neutron [-] [instance: 0704efaf-08de-4888-a092-476bcb789eea] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1235.975147] env[61987]: DEBUG nova.scheduler.client.report [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1236.155650] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062508, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488013} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.156017] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 054840f7-a1f2-4888-9a04-75c788dee325/054840f7-a1f2-4888-9a04-75c788dee325.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1236.156117] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1236.156376] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a3196348-13db-429c-b81b-993a3a4f706b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.162916] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1236.162916] env[61987]: value = "task-1062509" [ 1236.162916] env[61987]: _type = "Task" [ 1236.162916] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.171891] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062509, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.316114] env[61987]: DEBUG nova.compute.manager [req-76726be3-2e6a-4fb5-a938-f06ff08e2eb9 req-0b9c07be-918a-468f-b543-b4895c68fef4 service nova] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Received event network-vif-deleted-79e1d3b3-41db-4650-95bd-11df0dc06e80 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1236.316712] env[61987]: INFO nova.compute.manager [req-76726be3-2e6a-4fb5-a938-f06ff08e2eb9 req-0b9c07be-918a-468f-b543-b4895c68fef4 service nova] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Neutron deleted interface 79e1d3b3-41db-4650-95bd-11df0dc06e80; detaching it from the instance and deleting it from the info cache [ 1236.317026] env[61987]: DEBUG nova.network.neutron [req-76726be3-2e6a-4fb5-a938-f06ff08e2eb9 req-0b9c07be-918a-468f-b543-b4895c68fef4 service nova] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1236.363095] env[61987]: DEBUG nova.compute.manager [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1236.388712] env[61987]: DEBUG nova.virt.hardware [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1236.389050] env[61987]: DEBUG nova.virt.hardware [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1236.389246] env[61987]: DEBUG nova.virt.hardware [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1236.389444] env[61987]: DEBUG nova.virt.hardware [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1236.389598] env[61987]: DEBUG nova.virt.hardware [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1236.389750] env[61987]: DEBUG nova.virt.hardware [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1236.389972] env[61987]: DEBUG nova.virt.hardware [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1236.390391] env[61987]: DEBUG nova.virt.hardware [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1236.390391] env[61987]: DEBUG nova.virt.hardware [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1236.390553] env[61987]: DEBUG nova.virt.hardware [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1236.390703] env[61987]: DEBUG nova.virt.hardware [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1236.391818] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1693d06c-ba75-4d17-ac59-909ce14a07aa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.403286] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe68a14-4d86-4fc9-baf0-7c7c6645fdfe {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.417342] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Instance VIF info [] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1236.422925] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Creating folder: Project (14cc41dca5d74280ba207a810e77b60e). Parent ref: group-v234219. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1236.423186] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c671ffb5-1bfc-47cd-921c-2fb852171ad3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.433821] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Created folder: Project (14cc41dca5d74280ba207a810e77b60e) in parent group-v234219. [ 1236.434020] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Creating folder: Instances. Parent ref: group-v234459. {{(pid=61987) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1236.434250] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-438318af-8e8e-413d-bd2a-f56369c96850 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.442867] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Created folder: Instances in parent group-v234459. [ 1236.443323] env[61987]: DEBUG oslo.service.loopingcall [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1236.443415] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1236.443563] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ae9d5854-7a10-4f76-9b31-6a8b7f79c8ce {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.459584] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1236.459584] env[61987]: value = "task-1062512" [ 1236.459584] env[61987]: _type = "Task" [ 1236.459584] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.466841] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062512, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.479469] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.137s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1236.479976] env[61987]: DEBUG nova.compute.manager [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Start building networks asynchronously for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1236.674288] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062509, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063765} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.674628] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1236.675465] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8817ea05-f637-4efb-a86d-2e9c23461457 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.700468] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Reconfiguring VM instance instance-00000077 to attach disk [datastore2] 054840f7-a1f2-4888-9a04-75c788dee325/054840f7-a1f2-4888-9a04-75c788dee325.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1236.700889] env[61987]: DEBUG nova.network.neutron [-] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1236.702397] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c40cc47-55bf-4f6b-a2e0-e5e52c3aaa9e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.725058] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1236.725058] env[61987]: value = "task-1062513" [ 1236.725058] env[61987]: _type = "Task" [ 1236.725058] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.733344] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062513, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.820674] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-07832818-d5a7-4c85-a0a8-4ebad5ee037b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.829444] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13338ac4-e8f4-4263-9f40-30ceeb215cc1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.853180] env[61987]: DEBUG nova.compute.manager [req-76726be3-2e6a-4fb5-a938-f06ff08e2eb9 req-0b9c07be-918a-468f-b543-b4895c68fef4 service nova] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Detach interface failed, port_id=79e1d3b3-41db-4650-95bd-11df0dc06e80, reason: Instance 0704efaf-08de-4888-a092-476bcb789eea could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 1236.969763] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062512, 'name': CreateVM_Task, 'duration_secs': 0.337571} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.970120] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1236.970589] env[61987]: DEBUG oslo_concurrency.lockutils [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1236.970764] env[61987]: DEBUG oslo_concurrency.lockutils [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1236.971111] env[61987]: DEBUG oslo_concurrency.lockutils [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1236.971376] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a6304d3-574f-441e-9c5c-7d1d6b783594 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.975962] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1236.975962] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52dddc01-47de-1ebf-9b16-fd04d66ef9e8" [ 1236.975962] env[61987]: _type = "Task" [ 1236.975962] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.985508] env[61987]: DEBUG nova.compute.utils [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Using /dev/sd instead of None {{(pid=61987) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1236.986825] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52dddc01-47de-1ebf-9b16-fd04d66ef9e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.987136] env[61987]: DEBUG nova.compute.manager [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Not allocating networking since 'none' was specified. {{(pid=61987) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 1237.219983] env[61987]: INFO nova.compute.manager [-] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Took 1.52 seconds to deallocate network for instance. [ 1237.235023] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062513, 'name': ReconfigVM_Task, 'duration_secs': 0.363414} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.235159] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Reconfigured VM instance instance-00000077 to attach disk [datastore2] 054840f7-a1f2-4888-9a04-75c788dee325/054840f7-a1f2-4888-9a04-75c788dee325.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1237.235690] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-749af30e-3e45-47d3-b8bf-698f2b89ee59 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.241635] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1237.241635] env[61987]: value = "task-1062514" [ 1237.241635] env[61987]: _type = "Task" [ 1237.241635] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.249055] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062514, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.486619] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52dddc01-47de-1ebf-9b16-fd04d66ef9e8, 'name': SearchDatastore_Task, 'duration_secs': 0.008655} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.486897] env[61987]: DEBUG oslo_concurrency.lockutils [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1237.487164] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1237.487410] env[61987]: DEBUG oslo_concurrency.lockutils [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1237.487565] env[61987]: DEBUG oslo_concurrency.lockutils [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1237.487753] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1237.488025] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6d78bb3b-83ca-4a84-a596-14c8e9d86bee {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.491475] env[61987]: DEBUG nova.compute.manager [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Start building block device mappings for instance. {{(pid=61987) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1237.495598] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1237.495782] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1237.496486] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-225b88ea-978a-4ba5-9bf4-acf026994fc6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.501465] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1237.501465] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]523bf6df-aa0c-a638-8377-61dc00684b3e" [ 1237.501465] env[61987]: _type = "Task" [ 1237.501465] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.508701] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523bf6df-aa0c-a638-8377-61dc00684b3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.726682] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1237.726984] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1237.727230] env[61987]: DEBUG nova.objects.instance [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lazy-loading 'resources' on Instance uuid 0704efaf-08de-4888-a092-476bcb789eea {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1237.751061] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062514, 'name': Rename_Task, 'duration_secs': 0.135229} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.751292] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1237.751534] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bda4db65-2b15-4bff-939f-402ab19bd67e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.757291] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1237.757291] env[61987]: value = "task-1062515" [ 1237.757291] env[61987]: _type = "Task" [ 1237.757291] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.764373] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062515, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.013328] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]523bf6df-aa0c-a638-8377-61dc00684b3e, 'name': SearchDatastore_Task, 'duration_secs': 0.00762} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.014144] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52f81d8e-308f-4ca4-8f80-185ea8ab1b70 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.019523] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1238.019523] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]524dad1f-e5ae-f5d7-4762-48295fae5ca7" [ 1238.019523] env[61987]: _type = "Task" [ 1238.019523] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.027860] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524dad1f-e5ae-f5d7-4762-48295fae5ca7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.269081] env[61987]: DEBUG oslo_vmware.api [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062515, 'name': PowerOnVM_Task, 'duration_secs': 0.486181} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.269398] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1238.269561] env[61987]: INFO nova.compute.manager [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Took 6.47 seconds to spawn the instance on the hypervisor. [ 1238.269718] env[61987]: DEBUG nova.compute.manager [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1238.270549] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3505e9a0-c69a-43e2-abe9-05f42bd4c00b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.305561] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-939e04c6-3b8c-4819-80a1-2f8827aca976 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.314035] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9725004d-3d3e-43b2-91ac-1b5f5a59b930 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.345833] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8517adf-9832-450b-a05b-70e46ef63b11 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.353763] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f21aba9-1370-4aa9-96f1-b7f39b5b5cde {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.369649] env[61987]: DEBUG nova.compute.provider_tree [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1238.501799] env[61987]: DEBUG nova.compute.manager [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Start spawning the instance on the hypervisor. {{(pid=61987) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1238.526945] env[61987]: DEBUG nova.virt.hardware [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1238.527389] env[61987]: DEBUG nova.virt.hardware [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1238.527610] env[61987]: DEBUG nova.virt.hardware [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1238.528068] env[61987]: DEBUG nova.virt.hardware [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1238.528258] env[61987]: DEBUG nova.virt.hardware [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1238.528510] env[61987]: DEBUG nova.virt.hardware [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1238.528851] env[61987]: DEBUG nova.virt.hardware [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1238.529184] env[61987]: DEBUG nova.virt.hardware [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1238.529486] env[61987]: DEBUG nova.virt.hardware [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1238.529763] env[61987]: DEBUG nova.virt.hardware [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1238.530070] env[61987]: DEBUG nova.virt.hardware [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1238.531309] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5cf7d3-526f-436d-a4d0-59e725a3c48c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.547274] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e90e96c-2c61-45ea-bbb2-6a6117769a5e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.553256] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]524dad1f-e5ae-f5d7-4762-48295fae5ca7, 'name': SearchDatastore_Task, 'duration_secs': 0.010618} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.553689] env[61987]: DEBUG oslo_concurrency.lockutils [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1238.554112] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f/9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1238.555062] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-796c34de-0fdd-4caf-9fef-2ec2219f4a37 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.569196] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Instance VIF info [] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1238.574928] env[61987]: DEBUG oslo.service.loopingcall [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1238.575647] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1238.575889] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d47ad8ce-ef57-4289-859b-46b1ca75cf06 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.589608] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1238.589608] env[61987]: value = "task-1062516" [ 1238.589608] env[61987]: _type = "Task" [ 1238.589608] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.594928] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1238.594928] env[61987]: value = "task-1062517" [ 1238.594928] env[61987]: _type = "Task" [ 1238.594928] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.600905] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062516, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.605850] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062517, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.793784] env[61987]: INFO nova.compute.manager [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Took 11.17 seconds to build instance. [ 1238.872695] env[61987]: DEBUG nova.scheduler.client.report [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1239.107574] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062516, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.501784} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.111140] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f/9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1239.111380] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1239.111598] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062517, 'name': CreateVM_Task} progress is 99%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.111809] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7e345bc5-0600-4343-a6de-51f2a01b93dd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.117741] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1239.117741] env[61987]: value = "task-1062518" [ 1239.117741] env[61987]: _type = "Task" [ 1239.117741] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.125243] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062518, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.296763] env[61987]: DEBUG oslo_concurrency.lockutils [None req-a7314941-c1fa-4c46-b2d9-75f681d26bbb tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "054840f7-a1f2-4888-9a04-75c788dee325" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.680s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1239.378023] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.651s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1239.396325] env[61987]: INFO nova.scheduler.client.report [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Deleted allocations for instance 0704efaf-08de-4888-a092-476bcb789eea [ 1239.605128] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062517, 'name': CreateVM_Task, 'duration_secs': 0.519579} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.605356] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1239.605711] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1239.605881] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1239.606263] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1239.606531] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82b11cea-fba9-4e71-bf07-dc315f5b8fa5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.610752] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1239.610752] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52b2e9ed-e7ec-6869-8612-1b31ac62e19e" [ 1239.610752] env[61987]: _type = "Task" [ 1239.610752] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.619404] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b2e9ed-e7ec-6869-8612-1b31ac62e19e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.626136] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062518, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067872} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.626377] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1239.627087] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c87fb422-8d55-4056-95f6-12398fb56319 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.645745] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Reconfiguring VM instance instance-00000078 to attach disk [datastore2] 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f/9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1239.645929] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ca3e40c-6075-41ec-b5c8-2a6ac2de793b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.664786] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1239.664786] env[61987]: value = "task-1062519" [ 1239.664786] env[61987]: _type = "Task" [ 1239.664786] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.673569] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062519, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.904839] env[61987]: DEBUG oslo_concurrency.lockutils [None req-b8ad48d1-b694-4f2f-a57c-587c463e4b92 tempest-ServerActionsTestJSON-1504256817 tempest-ServerActionsTestJSON-1504256817-project-member] Lock "0704efaf-08de-4888-a092-476bcb789eea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.821s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1239.987385] env[61987]: DEBUG nova.compute.manager [req-cc6a3b56-3d56-4f80-bf7a-a184ff28832e req-8b89ad21-19fc-4763-a364-ade1534266fc service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Received event network-changed-7906310e-5bc1-4ad3-85fb-062c271a2f85 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1239.987557] env[61987]: DEBUG nova.compute.manager [req-cc6a3b56-3d56-4f80-bf7a-a184ff28832e req-8b89ad21-19fc-4763-a364-ade1534266fc service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Refreshing instance network info cache due to event network-changed-7906310e-5bc1-4ad3-85fb-062c271a2f85. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1239.987785] env[61987]: DEBUG oslo_concurrency.lockutils [req-cc6a3b56-3d56-4f80-bf7a-a184ff28832e req-8b89ad21-19fc-4763-a364-ade1534266fc service nova] Acquiring lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1239.987935] env[61987]: DEBUG oslo_concurrency.lockutils [req-cc6a3b56-3d56-4f80-bf7a-a184ff28832e req-8b89ad21-19fc-4763-a364-ade1534266fc service nova] Acquired lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1239.988138] env[61987]: DEBUG nova.network.neutron [req-cc6a3b56-3d56-4f80-bf7a-a184ff28832e req-8b89ad21-19fc-4763-a364-ade1534266fc service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Refreshing network info cache for port 7906310e-5bc1-4ad3-85fb-062c271a2f85 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1240.121487] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b2e9ed-e7ec-6869-8612-1b31ac62e19e, 'name': SearchDatastore_Task, 'duration_secs': 0.008292} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.121847] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1240.122106] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1240.122358] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1240.122517] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1240.122706] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1240.122961] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a69316d5-9145-4f07-8b73-5a1315a04f46 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.130778] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1240.130963] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1240.131655] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-289d4fdd-851a-4579-95dc-57f4e6288f89 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.136353] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1240.136353] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5201fd88-08ee-ac26-3c4a-c43e814319a2" [ 1240.136353] env[61987]: _type = "Task" [ 1240.136353] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.143416] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5201fd88-08ee-ac26-3c4a-c43e814319a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.173265] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062519, 'name': ReconfigVM_Task, 'duration_secs': 0.302422} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.173492] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Reconfigured VM instance instance-00000078 to attach disk [datastore2] 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f/9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1240.174071] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c973ce97-d035-4045-9ed3-62e622b56277 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.179895] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1240.179895] env[61987]: value = "task-1062520" [ 1240.179895] env[61987]: _type = "Task" [ 1240.179895] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.187830] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062520, 'name': Rename_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.646057] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]5201fd88-08ee-ac26-3c4a-c43e814319a2, 'name': SearchDatastore_Task, 'duration_secs': 0.007701} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.649105] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04848c94-1da3-4573-9e85-10140016d7ba {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.654685] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1240.654685] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52b2cc37-7436-6863-3c27-6d3e50e2a8a9" [ 1240.654685] env[61987]: _type = "Task" [ 1240.654685] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.661962] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b2cc37-7436-6863-3c27-6d3e50e2a8a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.690396] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062520, 'name': Rename_Task, 'duration_secs': 0.127758} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.690679] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1240.690934] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-04b1f2bc-d597-4e5c-973b-2c293d3b4039 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.697754] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1240.697754] env[61987]: value = "task-1062521" [ 1240.697754] env[61987]: _type = "Task" [ 1240.697754] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.706714] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062521, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.729432] env[61987]: DEBUG nova.network.neutron [req-cc6a3b56-3d56-4f80-bf7a-a184ff28832e req-8b89ad21-19fc-4763-a364-ade1534266fc service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Updated VIF entry in instance network info cache for port 7906310e-5bc1-4ad3-85fb-062c271a2f85. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1240.729810] env[61987]: DEBUG nova.network.neutron [req-cc6a3b56-3d56-4f80-bf7a-a184ff28832e req-8b89ad21-19fc-4763-a364-ade1534266fc service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Updating instance_info_cache with network_info: [{"id": "7906310e-5bc1-4ad3-85fb-062c271a2f85", "address": "fa:16:3e:99:f5:8a", "network": {"id": "eac55241-aad5-4ddb-a01e-625c9298a8af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1411727225-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90faa728219a432fa241b635071729e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f996252-e329-42bd-a897-446dfe2b81cd", "external-id": "nsx-vlan-transportzone-535", "segmentation_id": 535, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7906310e-5b", "ovs_interfaceid": "7906310e-5bc1-4ad3-85fb-062c271a2f85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1241.165904] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52b2cc37-7436-6863-3c27-6d3e50e2a8a9, 'name': SearchDatastore_Task, 'duration_secs': 0.008786} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.166205] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1241.166474] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 03766f52-c102-40ff-b004-7748231d0a03/03766f52-c102-40ff-b004-7748231d0a03.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1241.166739] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-35b15e61-89cb-4bf7-b588-11976318a1fb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.173633] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1241.173633] env[61987]: value = "task-1062522" [ 1241.173633] env[61987]: _type = "Task" [ 1241.173633] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1241.181884] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062522, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.206898] env[61987]: DEBUG oslo_vmware.api [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062521, 'name': PowerOnVM_Task, 'duration_secs': 0.40948} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.207142] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1241.207356] env[61987]: INFO nova.compute.manager [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Took 4.84 seconds to spawn the instance on the hypervisor. [ 1241.207543] env[61987]: DEBUG nova.compute.manager [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1241.208380] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b078d15d-2b6e-452c-80c4-33ef9c563188 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.232729] env[61987]: DEBUG oslo_concurrency.lockutils [req-cc6a3b56-3d56-4f80-bf7a-a184ff28832e req-8b89ad21-19fc-4763-a364-ade1534266fc service nova] Releasing lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1241.683759] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062522, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.726191] env[61987]: INFO nova.compute.manager [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Took 9.54 seconds to build instance. [ 1242.184330] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062522, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.526663} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.184614] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 03766f52-c102-40ff-b004-7748231d0a03/03766f52-c102-40ff-b004-7748231d0a03.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1242.184830] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1242.185094] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8376850c-8c86-4632-ab8f-7b7f93e0a7a0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.191631] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1242.191631] env[61987]: value = "task-1062523" [ 1242.191631] env[61987]: _type = "Task" [ 1242.191631] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.198888] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062523, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.231025] env[61987]: DEBUG oslo_concurrency.lockutils [None req-456a3e43-8b99-4b39-8132-535819934cb2 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.048s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1242.701440] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062523, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.389694} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.701922] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1242.702537] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f243d4a-75a7-4c8c-8be1-dc44a84e6d0b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.722798] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Reconfiguring VM instance instance-00000079 to attach disk [datastore2] 03766f52-c102-40ff-b004-7748231d0a03/03766f52-c102-40ff-b004-7748231d0a03.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1242.723094] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-286f77c5-80c2-46a0-bbd4-b5f7202b0512 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.742803] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1242.742803] env[61987]: value = "task-1062524" [ 1242.742803] env[61987]: _type = "Task" [ 1242.742803] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.750665] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062524, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.252500] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062524, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.753346] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062524, 'name': ReconfigVM_Task, 'duration_secs': 0.910497} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.753752] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Reconfigured VM instance instance-00000079 to attach disk [datastore2] 03766f52-c102-40ff-b004-7748231d0a03/03766f52-c102-40ff-b004-7748231d0a03.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1243.754244] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-96e54eea-57f4-449b-9853-c7e4fb901773 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.760631] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1243.760631] env[61987]: value = "task-1062525" [ 1243.760631] env[61987]: _type = "Task" [ 1243.760631] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.769093] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062525, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.270916] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062525, 'name': Rename_Task, 'duration_secs': 0.463954} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.271234] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1244.271493] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-127aae15-690d-4eb0-82cb-650809765df4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.277657] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1244.277657] env[61987]: value = "task-1062526" [ 1244.277657] env[61987]: _type = "Task" [ 1244.277657] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.284870] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062526, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.787682] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062526, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.288435] env[61987]: DEBUG oslo_vmware.api [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062526, 'name': PowerOnVM_Task, 'duration_secs': 0.587191} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.288706] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1245.288914] env[61987]: INFO nova.compute.manager [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Took 6.79 seconds to spawn the instance on the hypervisor. [ 1245.289155] env[61987]: DEBUG nova.compute.manager [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1245.289920] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a870fb2-cac2-496f-b53f-380f984fb540 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.805690] env[61987]: INFO nova.compute.manager [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Took 13.23 seconds to build instance. [ 1246.307717] env[61987]: DEBUG oslo_concurrency.lockutils [None req-ae96e407-b04c-4285-8d77-af1cd99c0e97 tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "03766f52-c102-40ff-b004-7748231d0a03" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.736s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1246.613881] env[61987]: INFO nova.compute.manager [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Rebuilding instance [ 1246.654975] env[61987]: DEBUG nova.compute.manager [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1246.655952] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ade7eb53-3b8a-4f12-970a-46d732758052 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.668372] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1247.668775] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7fcacdb4-8b84-4d4d-92c1-74ab8983c153 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.676272] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1247.676272] env[61987]: value = "task-1062527" [ 1247.676272] env[61987]: _type = "Task" [ 1247.676272] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.684378] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062527, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.186482] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062527, 'name': PowerOffVM_Task, 'duration_secs': 0.184541} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.186724] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1248.186952] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1248.187731] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6648967-b979-4cf3-95b8-a584b1fc8432 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.193996] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1248.194229] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a8f29af4-79b3-4162-8883-33d441b56dea {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.219030] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1248.219281] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1248.219472] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Deleting the datastore file [datastore2] 03766f52-c102-40ff-b004-7748231d0a03 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1248.219717] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0028ca82-708e-4d95-9270-97428d330f70 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.225694] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1248.225694] env[61987]: value = "task-1062529" [ 1248.225694] env[61987]: _type = "Task" [ 1248.225694] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1248.233069] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062529, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.735844] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062529, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.092357} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.736274] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1248.736321] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1248.736477] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1249.764515] env[61987]: DEBUG nova.virt.hardware [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-30T23:05:28Z,direct_url=,disk_format='vmdk',id=7bf8969f-d179-42af-9e15-527d5449f929,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='705320a8b6a549d0b197ee1a35404d52',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-30T23:05:28Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1249.764816] env[61987]: DEBUG nova.virt.hardware [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1249.764966] env[61987]: DEBUG nova.virt.hardware [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1249.765179] env[61987]: DEBUG nova.virt.hardware [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1249.765339] env[61987]: DEBUG nova.virt.hardware [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1249.765499] env[61987]: DEBUG nova.virt.hardware [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1249.765715] env[61987]: DEBUG nova.virt.hardware [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1249.765885] env[61987]: DEBUG nova.virt.hardware [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1249.766074] env[61987]: DEBUG nova.virt.hardware [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1249.766255] env[61987]: DEBUG nova.virt.hardware [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1249.766440] env[61987]: DEBUG nova.virt.hardware [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1249.767378] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dea81f9-1704-4cad-a294-2873b25fcd3b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.775246] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bdf4556-eba8-44ef-9d7a-3946326a6c86 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.788575] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Instance VIF info [] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1249.794143] env[61987]: DEBUG oslo.service.loopingcall [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1249.794377] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1249.794584] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ac3cf2d7-22db-4de8-87cd-56d68208430e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.810608] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1249.810608] env[61987]: value = "task-1062530" [ 1249.810608] env[61987]: _type = "Task" [ 1249.810608] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.817657] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062530, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.320770] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062530, 'name': CreateVM_Task, 'duration_secs': 0.250536} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.320960] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1250.321521] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1250.321797] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1250.322314] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1250.322673] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37757d93-7f35-48fa-a49d-b193f9656cd4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.326902] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1250.326902] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52d7f287-762c-6f37-6316-57f9e28f3e1b" [ 1250.326902] env[61987]: _type = "Task" [ 1250.326902] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.334054] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52d7f287-762c-6f37-6316-57f9e28f3e1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.837045] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52d7f287-762c-6f37-6316-57f9e28f3e1b, 'name': SearchDatastore_Task, 'duration_secs': 0.00872} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.837422] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1250.837636] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Processing image 7bf8969f-d179-42af-9e15-527d5449f929 {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1250.837882] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1250.838051] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquired lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1250.838243] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1250.838517] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3516caa6-8312-4674-ac4a-effcde63c401 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.845915] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1250.846103] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1250.846784] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b476ce1-4bb0-433e-aadf-78aca5b54aef {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.851367] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1250.851367] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52c346ff-a73d-e023-f404-bdeca9affbcd" [ 1250.851367] env[61987]: _type = "Task" [ 1250.851367] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.858108] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c346ff-a73d-e023-f404-bdeca9affbcd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.361725] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52c346ff-a73d-e023-f404-bdeca9affbcd, 'name': SearchDatastore_Task, 'duration_secs': 0.007905} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.362482] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bdf4a053-3d34-4a00-876d-1a844f528689 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.367517] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1251.367517] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52e2f965-6397-b18c-cdf9-8780ef348652" [ 1251.367517] env[61987]: _type = "Task" [ 1251.367517] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.375349] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52e2f965-6397-b18c-cdf9-8780ef348652, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.879905] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52e2f965-6397-b18c-cdf9-8780ef348652, 'name': SearchDatastore_Task, 'duration_secs': 0.015176} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.880315] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Releasing lock "[datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1251.880494] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 03766f52-c102-40ff-b004-7748231d0a03/03766f52-c102-40ff-b004-7748231d0a03.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1251.880750] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-56ecd131-eb70-4217-b478-c36fba70210b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.886983] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1251.886983] env[61987]: value = "task-1062531" [ 1251.886983] env[61987]: _type = "Task" [ 1251.886983] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.893994] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062531, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.397222] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062531, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.424509} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.397493] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/7bf8969f-d179-42af-9e15-527d5449f929/7bf8969f-d179-42af-9e15-527d5449f929.vmdk to [datastore2] 03766f52-c102-40ff-b004-7748231d0a03/03766f52-c102-40ff-b004-7748231d0a03.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1252.397707] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Extending root virtual disk to 1048576 {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1252.397955] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b0794f21-9f27-455e-83e6-76ed3ef96ce6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.405077] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1252.405077] env[61987]: value = "task-1062532" [ 1252.405077] env[61987]: _type = "Task" [ 1252.405077] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.411790] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062532, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.915798] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062532, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059147} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.916204] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Extended root virtual disk {{(pid=61987) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1252.916879] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-670c0fcb-115d-4bbd-997a-fcd6f7e71f6d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.936576] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Reconfiguring VM instance instance-00000079 to attach disk [datastore2] 03766f52-c102-40ff-b004-7748231d0a03/03766f52-c102-40ff-b004-7748231d0a03.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1252.936826] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f04f7da-34ca-46f2-af26-35fa571ff3ce {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.955770] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1252.955770] env[61987]: value = "task-1062533" [ 1252.955770] env[61987]: _type = "Task" [ 1252.955770] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.963213] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062533, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.465500] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062533, 'name': ReconfigVM_Task, 'duration_secs': 0.254336} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.465805] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Reconfigured VM instance instance-00000079 to attach disk [datastore2] 03766f52-c102-40ff-b004-7748231d0a03/03766f52-c102-40ff-b004-7748231d0a03.vmdk or device None with type sparse {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1253.466439] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6afbc77a-0c91-4c62-86b8-280373ca3587 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.472957] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1253.472957] env[61987]: value = "task-1062534" [ 1253.472957] env[61987]: _type = "Task" [ 1253.472957] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.480335] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062534, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.982691] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062534, 'name': Rename_Task, 'duration_secs': 0.126211} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.983098] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1253.983234] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-442c9122-be80-478d-9834-c556837173be {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.989917] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1253.989917] env[61987]: value = "task-1062535" [ 1253.989917] env[61987]: _type = "Task" [ 1253.989917] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.997268] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062535, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.499603] env[61987]: DEBUG oslo_vmware.api [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062535, 'name': PowerOnVM_Task, 'duration_secs': 0.380461} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.499949] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1254.500185] env[61987]: DEBUG nova.compute.manager [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1254.500932] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-433eeed2-ccfc-46c8-88dc-3cbc36cbf808 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.019932] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1255.020319] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1255.020571] env[61987]: DEBUG nova.objects.instance [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61987) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1255.397903] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquiring lock "03766f52-c102-40ff-b004-7748231d0a03" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1255.398215] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "03766f52-c102-40ff-b004-7748231d0a03" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1255.398436] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquiring lock "03766f52-c102-40ff-b004-7748231d0a03-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1255.398624] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "03766f52-c102-40ff-b004-7748231d0a03-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1255.398804] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "03766f52-c102-40ff-b004-7748231d0a03-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1255.400998] env[61987]: INFO nova.compute.manager [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Terminating instance [ 1255.904176] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquiring lock "refresh_cache-03766f52-c102-40ff-b004-7748231d0a03" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1255.904456] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquired lock "refresh_cache-03766f52-c102-40ff-b004-7748231d0a03" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1255.904583] env[61987]: DEBUG nova.network.neutron [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1256.031503] env[61987]: DEBUG oslo_concurrency.lockutils [None req-4ecbbcda-1a7f-44d3-8b99-b6ec68303caf tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1256.422992] env[61987]: DEBUG nova.network.neutron [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1256.475287] env[61987]: DEBUG nova.network.neutron [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1256.965276] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1256.978136] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Releasing lock "refresh_cache-03766f52-c102-40ff-b004-7748231d0a03" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1256.978590] env[61987]: DEBUG nova.compute.manager [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1256.978790] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1256.979752] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aa6b829-a52f-4299-9fa9-b2208a0fce72 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.987501] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1256.987743] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b8cbe623-51e3-47cf-b2c3-c71cd9849412 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.993543] env[61987]: DEBUG oslo_vmware.api [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1256.993543] env[61987]: value = "task-1062536" [ 1256.993543] env[61987]: _type = "Task" [ 1256.993543] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.001189] env[61987]: DEBUG oslo_vmware.api [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062536, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.503663] env[61987]: DEBUG oslo_vmware.api [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062536, 'name': PowerOffVM_Task, 'duration_secs': 0.17287} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.504046] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1257.504117] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1257.504373] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6c586fc7-baff-4b82-9a63-c3b330ed2e3d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.526216] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1257.526463] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1257.526657] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Deleting the datastore file [datastore2] 03766f52-c102-40ff-b004-7748231d0a03 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1257.526906] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a83a1f6f-eb89-47b7-8aac-67b8d31282a2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.532937] env[61987]: DEBUG oslo_vmware.api [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1257.532937] env[61987]: value = "task-1062538" [ 1257.532937] env[61987]: _type = "Task" [ 1257.532937] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.540404] env[61987]: DEBUG oslo_vmware.api [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062538, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.965558] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1257.965756] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61987) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11020}} [ 1258.043073] env[61987]: DEBUG oslo_vmware.api [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062538, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.087154} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.043356] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1258.043553] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1258.043739] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1258.043922] env[61987]: INFO nova.compute.manager [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Took 1.07 seconds to destroy the instance on the hypervisor. [ 1258.044189] env[61987]: DEBUG oslo.service.loopingcall [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1258.044390] env[61987]: DEBUG nova.compute.manager [-] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1258.044487] env[61987]: DEBUG nova.network.neutron [-] [instance: 03766f52-c102-40ff-b004-7748231d0a03] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1258.058979] env[61987]: DEBUG nova.network.neutron [-] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1258.561689] env[61987]: DEBUG nova.network.neutron [-] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1258.966126] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1259.064620] env[61987]: INFO nova.compute.manager [-] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Took 1.02 seconds to deallocate network for instance. [ 1259.570646] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1259.571065] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1259.571227] env[61987]: DEBUG nova.objects.instance [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lazy-loading 'resources' on Instance uuid 03766f52-c102-40ff-b004-7748231d0a03 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1259.960758] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1259.965375] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1259.965541] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Starting heal instance info cache {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10401}} [ 1260.122299] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bbaf323-ac42-4713-a80b-3eeb63589ada {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.129466] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beacaacd-15c7-4954-b7f2-414680083227 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.158557] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c3c92a5-7a15-42de-90db-47b5d2fa77d1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.165084] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0be825-a9ea-40dc-8056-201645c960fd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.177416] env[61987]: DEBUG nova.compute.provider_tree [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1260.697346] env[61987]: ERROR nova.scheduler.client.report [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [req-0da28f70-8242-47d0-b054-7049aad4aa6f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8db744e5-1e14-4e98-9818-695ea4b7dfd0. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-0da28f70-8242-47d0-b054-7049aad4aa6f"}]} [ 1260.712908] env[61987]: DEBUG nova.scheduler.client.report [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Refreshing inventories for resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1260.724930] env[61987]: DEBUG nova.scheduler.client.report [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Updating ProviderTree inventory for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1260.725164] env[61987]: DEBUG nova.compute.provider_tree [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1260.734707] env[61987]: DEBUG nova.scheduler.client.report [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Refreshing aggregate associations for resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0, aggregates: None {{(pid=61987) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1260.752381] env[61987]: DEBUG nova.scheduler.client.report [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Refreshing trait associations for resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61987) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1260.791599] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc142e10-5395-4141-9653-13f14245044e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.799057] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26856e9a-6575-4e93-9c19-f3c92eb3f190 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.829225] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01456339-47dd-4bf7-8af8-ab9cc8724e7a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.836506] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de20b617-fdcf-4f41-97be-b4ba21a2902e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.849492] env[61987]: DEBUG nova.compute.provider_tree [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1261.378829] env[61987]: DEBUG nova.scheduler.client.report [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Updated inventory for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with generation 155 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1261.379128] env[61987]: DEBUG nova.compute.provider_tree [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Updating resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 generation from 155 to 156 during operation: update_inventory {{(pid=61987) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1261.379381] env[61987]: DEBUG nova.compute.provider_tree [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1261.884370] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.313s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1261.904748] env[61987]: INFO nova.scheduler.client.report [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Deleted allocations for instance 03766f52-c102-40ff-b004-7748231d0a03 [ 1261.974793] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Didn't find any instances for network info cache update. {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10487}} [ 1261.975019] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1261.975199] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1261.975349] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1261.975503] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1262.414819] env[61987]: DEBUG oslo_concurrency.lockutils [None req-0d9cbbcd-8770-4894-93ad-ae027829f2be tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "03766f52-c102-40ff-b004-7748231d0a03" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.017s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1262.479059] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1262.479059] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1262.479059] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1262.479263] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61987) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1262.480068] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-660d3f4e-73cf-4ea2-8ced-85c51f454cbc {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.489765] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c8dd3a-7987-4d67-8b71-e7a8591522d0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.503609] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d6cb30c-08d0-425d-abe8-78e40d917b52 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.509531] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4340665-25cc-436c-87b8-850fddc02e85 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.537915] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181513MB free_disk=179GB free_vcpus=48 pci_devices=None {{(pid=61987) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1262.538072] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1262.538289] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1262.906338] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquiring lock "9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1262.906774] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1262.906816] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquiring lock "9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1262.907016] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1262.907228] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1262.909878] env[61987]: INFO nova.compute.manager [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Terminating instance [ 1263.413648] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquiring lock "refresh_cache-9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1263.413861] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquired lock "refresh_cache-9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1263.414030] env[61987]: DEBUG nova.network.neutron [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1263.561337] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 054840f7-a1f2-4888-9a04-75c788dee325 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1263.561503] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1263.561689] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1263.561866] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1263.594970] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c12fc640-69bd-4325-9307-4f30ac705e72 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.602679] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f886aa-d2ad-4a24-a7af-d2b8b96382eb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.632253] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc98f9e2-881d-4d95-b67a-a3e2afbe44cb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.639927] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1846cd22-93d8-4a7b-a612-ace411f6a7f8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.652710] env[61987]: DEBUG nova.compute.provider_tree [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1263.933358] env[61987]: DEBUG nova.network.neutron [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1263.987693] env[61987]: DEBUG nova.network.neutron [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1264.155165] env[61987]: DEBUG nova.scheduler.client.report [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1264.492140] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Releasing lock "refresh_cache-9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1264.492140] env[61987]: DEBUG nova.compute.manager [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1264.492140] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1264.492895] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c14d2e8-1bfc-463a-a5b7-415e88cdf85b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.500342] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1264.500583] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-37409944-b86d-4546-99ad-cd5f53d4766e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.507165] env[61987]: DEBUG oslo_vmware.api [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1264.507165] env[61987]: value = "task-1062539" [ 1264.507165] env[61987]: _type = "Task" [ 1264.507165] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.515526] env[61987]: DEBUG oslo_vmware.api [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062539, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.659632] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61987) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1264.659840] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.122s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1265.017075] env[61987]: DEBUG oslo_vmware.api [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062539, 'name': PowerOffVM_Task, 'duration_secs': 0.114633} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.017478] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1265.017529] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1265.017765] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d3829bfe-6ae1-46be-b44e-c6129a996e0a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.042043] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1265.042259] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1265.042448] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Deleting the datastore file [datastore2] 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1265.042695] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b5427abb-b0ee-422d-98a4-7f1dce7f868f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.048937] env[61987]: DEBUG oslo_vmware.api [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for the task: (returnval){ [ 1265.048937] env[61987]: value = "task-1062541" [ 1265.048937] env[61987]: _type = "Task" [ 1265.048937] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.055934] env[61987]: DEBUG oslo_vmware.api [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062541, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.559077] env[61987]: DEBUG oslo_vmware.api [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Task: {'id': task-1062541, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.085393} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.559369] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1265.559574] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1265.559768] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1265.559945] env[61987]: INFO nova.compute.manager [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Took 1.07 seconds to destroy the instance on the hypervisor. [ 1265.560219] env[61987]: DEBUG oslo.service.loopingcall [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1265.560435] env[61987]: DEBUG nova.compute.manager [-] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1265.560530] env[61987]: DEBUG nova.network.neutron [-] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1265.575182] env[61987]: DEBUG nova.network.neutron [-] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Instance cache missing network info. {{(pid=61987) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1266.078181] env[61987]: DEBUG nova.network.neutron [-] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1266.581146] env[61987]: INFO nova.compute.manager [-] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Took 1.02 seconds to deallocate network for instance. [ 1267.781606] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1267.781977] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1267.782257] env[61987]: DEBUG nova.objects.instance [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lazy-loading 'resources' on Instance uuid 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1268.320493] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a45ffcc2-b992-4f25-8405-4ab467acd6a7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.328158] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-192428a6-b034-4b74-8f9d-8aa941029271 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.357214] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1820f531-cf4d-48a7-89b2-3f3278d554b8 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.364725] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde85567-a063-4953-9e06-6f04178e1fcb {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.377179] env[61987]: DEBUG nova.compute.provider_tree [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1268.907067] env[61987]: DEBUG nova.scheduler.client.report [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Updated inventory for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with generation 156 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1268.907382] env[61987]: DEBUG nova.compute.provider_tree [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Updating resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 generation from 156 to 157 during operation: update_inventory {{(pid=61987) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1268.907587] env[61987]: DEBUG nova.compute.provider_tree [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1269.412562] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.630s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1269.431027] env[61987]: INFO nova.scheduler.client.report [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Deleted allocations for instance 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f [ 1269.938512] env[61987]: DEBUG oslo_concurrency.lockutils [None req-2c5fb148-12b8-4d7e-a12a-321acb4e566e tempest-ServerShowV247Test-1825420699 tempest-ServerShowV247Test-1825420699-project-member] Lock "9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.032s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1279.166721] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "054840f7-a1f2-4888-9a04-75c788dee325" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1279.167153] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "054840f7-a1f2-4888-9a04-75c788dee325" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1279.167264] env[61987]: INFO nova.compute.manager [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Shelving [ 1280.177084] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1280.178021] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bd54cb77-e14c-47f6-a59b-2ec81955fdc0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.184955] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1280.184955] env[61987]: value = "task-1062542" [ 1280.184955] env[61987]: _type = "Task" [ 1280.184955] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1280.192654] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062542, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1280.696384] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062542, 'name': PowerOffVM_Task, 'duration_secs': 0.172632} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.696798] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1280.697680] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa7399fb-3a50-4037-aba7-85d352da5486 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1280.718665] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9423a575-36d2-4909-873a-3e1f11914698 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.229577] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Creating Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1281.229930] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-1fd07480-a153-409d-90e4-e0685bd1c2b4 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1281.238729] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1281.238729] env[61987]: value = "task-1062543" [ 1281.238729] env[61987]: _type = "Task" [ 1281.238729] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1281.246655] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062543, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1281.748445] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062543, 'name': CreateSnapshot_Task, 'duration_secs': 0.378281} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1281.748702] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Created Snapshot of the VM instance {{(pid=61987) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1281.749425] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d00e5c3f-7e8b-48fd-8157-a902e7c35f01 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.265494] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Creating linked-clone VM from snapshot {{(pid=61987) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1282.265848] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8ce5e0c7-f298-491a-96e0-dfa83d25d678 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.274168] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1282.274168] env[61987]: value = "task-1062544" [ 1282.274168] env[61987]: _type = "Task" [ 1282.274168] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1282.281958] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062544, 'name': CloneVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1282.783860] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062544, 'name': CloneVM_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1283.284029] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062544, 'name': CloneVM_Task, 'duration_secs': 0.866088} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1283.287030] env[61987]: INFO nova.virt.vmwareapi.vmops [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Created linked-clone VM from snapshot [ 1283.287030] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d31a366-a4e4-4fd3-9c26-50ca32f9894b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.292083] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Uploading image e6a4516f-aab7-4816-a4c8-1b5f531bee5b {{(pid=61987) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1283.311366] env[61987]: DEBUG oslo_vmware.rw_handles [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1283.311366] env[61987]: value = "vm-234465" [ 1283.311366] env[61987]: _type = "VirtualMachine" [ 1283.311366] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1283.311592] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d52d5c75-6700-4c67-8dbb-ea0f8fb3f597 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.317199] env[61987]: DEBUG oslo_vmware.rw_handles [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lease: (returnval){ [ 1283.317199] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52994452-4e95-a403-cd07-ce3a4f749597" [ 1283.317199] env[61987]: _type = "HttpNfcLease" [ 1283.317199] env[61987]: } obtained for exporting VM: (result){ [ 1283.317199] env[61987]: value = "vm-234465" [ 1283.317199] env[61987]: _type = "VirtualMachine" [ 1283.317199] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1283.317442] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the lease: (returnval){ [ 1283.317442] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52994452-4e95-a403-cd07-ce3a4f749597" [ 1283.317442] env[61987]: _type = "HttpNfcLease" [ 1283.317442] env[61987]: } to be ready. {{(pid=61987) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1283.322927] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1283.322927] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52994452-4e95-a403-cd07-ce3a4f749597" [ 1283.322927] env[61987]: _type = "HttpNfcLease" [ 1283.322927] env[61987]: } is initializing. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1283.825719] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1283.825719] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52994452-4e95-a403-cd07-ce3a4f749597" [ 1283.825719] env[61987]: _type = "HttpNfcLease" [ 1283.825719] env[61987]: } is ready. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1283.826145] env[61987]: DEBUG oslo_vmware.rw_handles [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1283.826145] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52994452-4e95-a403-cd07-ce3a4f749597" [ 1283.826145] env[61987]: _type = "HttpNfcLease" [ 1283.826145] env[61987]: }. {{(pid=61987) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1283.826760] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f996714c-e371-437b-9840-253f2e9da3e6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.833305] env[61987]: DEBUG oslo_vmware.rw_handles [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc1ac9-568f-e3bf-0b75-f764d52accb4/disk-0.vmdk from lease info. {{(pid=61987) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1283.833483] env[61987]: DEBUG oslo_vmware.rw_handles [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc1ac9-568f-e3bf-0b75-f764d52accb4/disk-0.vmdk for reading. {{(pid=61987) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1283.920950] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0d9e2fee-05be-449e-912a-efb7037b2cf1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.279308] env[61987]: DEBUG oslo_vmware.rw_handles [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc1ac9-568f-e3bf-0b75-f764d52accb4/disk-0.vmdk. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1291.280388] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20a8c445-dae6-452d-98c4-1df35f466786 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.286569] env[61987]: DEBUG oslo_vmware.rw_handles [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc1ac9-568f-e3bf-0b75-f764d52accb4/disk-0.vmdk is in state: ready. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1291.286743] env[61987]: ERROR oslo_vmware.rw_handles [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc1ac9-568f-e3bf-0b75-f764d52accb4/disk-0.vmdk due to incomplete transfer. [ 1291.286967] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-8f46e14e-cd79-4132-ac6a-3b83025ef72d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.293858] env[61987]: DEBUG oslo_vmware.rw_handles [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc1ac9-568f-e3bf-0b75-f764d52accb4/disk-0.vmdk. {{(pid=61987) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1291.294096] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Uploaded image e6a4516f-aab7-4816-a4c8-1b5f531bee5b to the Glance image server {{(pid=61987) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1291.296304] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Destroying the VM {{(pid=61987) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1291.296540] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f7765ff2-015b-4bf9-904a-64a40767be31 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.302184] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1291.302184] env[61987]: value = "task-1062546" [ 1291.302184] env[61987]: _type = "Task" [ 1291.302184] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1291.309369] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062546, 'name': Destroy_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1291.812075] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062546, 'name': Destroy_Task, 'duration_secs': 0.288512} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1291.812351] env[61987]: INFO nova.virt.vmwareapi.vm_util [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Destroyed the VM [ 1291.812590] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Deleting Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1291.812840] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0b166c37-f16d-47e3-be7c-c37db26634c9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.818769] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1291.818769] env[61987]: value = "task-1062547" [ 1291.818769] env[61987]: _type = "Task" [ 1291.818769] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1291.825842] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062547, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.328632] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062547, 'name': RemoveSnapshot_Task, 'duration_secs': 0.363721} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1292.328948] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Deleted Snapshot of the VM instance {{(pid=61987) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1292.329218] env[61987]: DEBUG nova.compute.manager [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1292.330050] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27dca823-0d9e-4ab5-86e4-3f1792fb92a3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.841521] env[61987]: INFO nova.compute.manager [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Shelve offloading [ 1293.345866] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1293.346546] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-12a8043b-4dea-4d58-a53b-5b4f689027d6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.355026] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1293.355026] env[61987]: value = "task-1062548" [ 1293.355026] env[61987]: _type = "Task" [ 1293.355026] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1293.362576] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062548, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.865394] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] VM already powered off {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1293.865619] env[61987]: DEBUG nova.compute.manager [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1293.866390] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba34368b-967b-4155-8cf8-0e0c9ffebd91 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1293.871847] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1293.872035] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquired lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1293.872218] env[61987]: DEBUG nova.network.neutron [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1294.620150] env[61987]: DEBUG nova.network.neutron [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Updating instance_info_cache with network_info: [{"id": "7906310e-5bc1-4ad3-85fb-062c271a2f85", "address": "fa:16:3e:99:f5:8a", "network": {"id": "eac55241-aad5-4ddb-a01e-625c9298a8af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1411727225-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90faa728219a432fa241b635071729e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f996252-e329-42bd-a897-446dfe2b81cd", "external-id": "nsx-vlan-transportzone-535", "segmentation_id": 535, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7906310e-5b", "ovs_interfaceid": "7906310e-5bc1-4ad3-85fb-062c271a2f85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1295.123372] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Releasing lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1295.331343] env[61987]: DEBUG nova.compute.manager [req-efbd9dae-8df0-4603-a9f3-f8d9f3bd5f11 req-4bb2f7ab-0351-468c-9ff1-84138158d833 service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Received event network-vif-unplugged-7906310e-5bc1-4ad3-85fb-062c271a2f85 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1295.331594] env[61987]: DEBUG oslo_concurrency.lockutils [req-efbd9dae-8df0-4603-a9f3-f8d9f3bd5f11 req-4bb2f7ab-0351-468c-9ff1-84138158d833 service nova] Acquiring lock "054840f7-a1f2-4888-9a04-75c788dee325-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1295.331802] env[61987]: DEBUG oslo_concurrency.lockutils [req-efbd9dae-8df0-4603-a9f3-f8d9f3bd5f11 req-4bb2f7ab-0351-468c-9ff1-84138158d833 service nova] Lock "054840f7-a1f2-4888-9a04-75c788dee325-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1295.331978] env[61987]: DEBUG oslo_concurrency.lockutils [req-efbd9dae-8df0-4603-a9f3-f8d9f3bd5f11 req-4bb2f7ab-0351-468c-9ff1-84138158d833 service nova] Lock "054840f7-a1f2-4888-9a04-75c788dee325-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1295.332245] env[61987]: DEBUG nova.compute.manager [req-efbd9dae-8df0-4603-a9f3-f8d9f3bd5f11 req-4bb2f7ab-0351-468c-9ff1-84138158d833 service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] No waiting events found dispatching network-vif-unplugged-7906310e-5bc1-4ad3-85fb-062c271a2f85 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1295.332432] env[61987]: WARNING nova.compute.manager [req-efbd9dae-8df0-4603-a9f3-f8d9f3bd5f11 req-4bb2f7ab-0351-468c-9ff1-84138158d833 service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Received unexpected event network-vif-unplugged-7906310e-5bc1-4ad3-85fb-062c271a2f85 for instance with vm_state shelved and task_state shelving_offloading. [ 1295.416649] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1295.417543] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a890a7ec-cf34-4749-b935-6b439f56090e {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.424997] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1295.425240] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9407b777-b05c-4c42-aa48-d3870711bf95 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.488297] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1295.488545] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1295.488735] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Deleting the datastore file [datastore2] 054840f7-a1f2-4888-9a04-75c788dee325 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1295.489016] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2a8278e3-ef16-4c7f-93dd-933e0c4047b1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1295.495735] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1295.495735] env[61987]: value = "task-1062550" [ 1295.495735] env[61987]: _type = "Task" [ 1295.495735] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1295.502843] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062550, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1296.005276] env[61987]: DEBUG oslo_vmware.api [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062550, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130821} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1296.005634] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1296.005678] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1296.005837] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1296.029957] env[61987]: INFO nova.scheduler.client.report [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Deleted allocations for instance 054840f7-a1f2-4888-9a04-75c788dee325 [ 1296.535033] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1296.535273] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1296.535500] env[61987]: DEBUG nova.objects.instance [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lazy-loading 'resources' on Instance uuid 054840f7-a1f2-4888-9a04-75c788dee325 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1297.040552] env[61987]: DEBUG nova.objects.instance [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lazy-loading 'numa_topology' on Instance uuid 054840f7-a1f2-4888-9a04-75c788dee325 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1297.355892] env[61987]: DEBUG nova.compute.manager [req-d861be57-ff8a-418e-8ac1-61e5350f2bbf req-941a9555-414b-4042-ad46-9d147210312b service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Received event network-changed-7906310e-5bc1-4ad3-85fb-062c271a2f85 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1297.356048] env[61987]: DEBUG nova.compute.manager [req-d861be57-ff8a-418e-8ac1-61e5350f2bbf req-941a9555-414b-4042-ad46-9d147210312b service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Refreshing instance network info cache due to event network-changed-7906310e-5bc1-4ad3-85fb-062c271a2f85. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1297.356284] env[61987]: DEBUG oslo_concurrency.lockutils [req-d861be57-ff8a-418e-8ac1-61e5350f2bbf req-941a9555-414b-4042-ad46-9d147210312b service nova] Acquiring lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1297.356435] env[61987]: DEBUG oslo_concurrency.lockutils [req-d861be57-ff8a-418e-8ac1-61e5350f2bbf req-941a9555-414b-4042-ad46-9d147210312b service nova] Acquired lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1297.356599] env[61987]: DEBUG nova.network.neutron [req-d861be57-ff8a-418e-8ac1-61e5350f2bbf req-941a9555-414b-4042-ad46-9d147210312b service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Refreshing network info cache for port 7906310e-5bc1-4ad3-85fb-062c271a2f85 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1297.542548] env[61987]: DEBUG nova.objects.base [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Object Instance<054840f7-a1f2-4888-9a04-75c788dee325> lazy-loaded attributes: resources,numa_topology {{(pid=61987) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1297.556498] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c86abfaf-20ca-4365-b938-79c276d68ea5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.564046] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a39fb491-0c52-4a5d-b38a-173d948291c6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.592456] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-636807f0-638b-40df-a33d-d5031ee3cead {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.599044] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46403056-6c3a-47eb-b68b-eb175785fb6b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.611949] env[61987]: DEBUG nova.compute.provider_tree [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1298.085270] env[61987]: DEBUG nova.network.neutron [req-d861be57-ff8a-418e-8ac1-61e5350f2bbf req-941a9555-414b-4042-ad46-9d147210312b service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Updated VIF entry in instance network info cache for port 7906310e-5bc1-4ad3-85fb-062c271a2f85. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1298.085731] env[61987]: DEBUG nova.network.neutron [req-d861be57-ff8a-418e-8ac1-61e5350f2bbf req-941a9555-414b-4042-ad46-9d147210312b service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Updating instance_info_cache with network_info: [{"id": "7906310e-5bc1-4ad3-85fb-062c271a2f85", "address": "fa:16:3e:99:f5:8a", "network": {"id": "eac55241-aad5-4ddb-a01e-625c9298a8af", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1411727225-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90faa728219a432fa241b635071729e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap7906310e-5b", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1298.114872] env[61987]: DEBUG nova.scheduler.client.report [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1298.588787] env[61987]: DEBUG oslo_concurrency.lockutils [req-d861be57-ff8a-418e-8ac1-61e5350f2bbf req-941a9555-414b-4042-ad46-9d147210312b service nova] Releasing lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1298.620619] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.085s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1298.982805] env[61987]: DEBUG oslo_concurrency.lockutils [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "054840f7-a1f2-4888-9a04-75c788dee325" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1299.131260] env[61987]: DEBUG oslo_concurrency.lockutils [None req-6fc69d24-8230-416f-b8a1-05b02db69b37 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "054840f7-a1f2-4888-9a04-75c788dee325" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 19.964s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1299.132571] env[61987]: DEBUG oslo_concurrency.lockutils [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "054840f7-a1f2-4888-9a04-75c788dee325" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.150s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1299.132854] env[61987]: INFO nova.compute.manager [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Unshelving [ 1300.154208] env[61987]: DEBUG oslo_concurrency.lockutils [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1300.154505] env[61987]: DEBUG oslo_concurrency.lockutils [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1300.154715] env[61987]: DEBUG nova.objects.instance [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lazy-loading 'pci_requests' on Instance uuid 054840f7-a1f2-4888-9a04-75c788dee325 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1300.658438] env[61987]: DEBUG nova.objects.instance [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lazy-loading 'numa_topology' on Instance uuid 054840f7-a1f2-4888-9a04-75c788dee325 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1301.161550] env[61987]: INFO nova.compute.claims [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1302.199534] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1400e6e8-fae3-41bb-8669-0784eb5f0ef2 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.206827] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af47b37-db05-4196-97a7-7f4cecf4e5b0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.236279] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e58f23d9-8f52-446c-8ea1-2dd4dbad91fa {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.243183] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb2fd56-6b6b-4ff3-bdf6-e724c5b8333c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.255640] env[61987]: DEBUG nova.compute.provider_tree [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Inventory has not changed in ProviderTree for provider: 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1302.760425] env[61987]: DEBUG nova.scheduler.client.report [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Inventory has not changed for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1303.265828] env[61987]: DEBUG oslo_concurrency.lockutils [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.111s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1303.294340] env[61987]: INFO nova.network.neutron [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Updating port 7906310e-5bc1-4ad3-85fb-062c271a2f85 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1304.715859] env[61987]: DEBUG nova.compute.manager [req-a81410f9-ae3a-4248-a4a6-ebe6a756b96e req-4f48af90-823d-4250-9836-5dd843722eef service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Received event network-vif-plugged-7906310e-5bc1-4ad3-85fb-062c271a2f85 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1304.717176] env[61987]: DEBUG oslo_concurrency.lockutils [req-a81410f9-ae3a-4248-a4a6-ebe6a756b96e req-4f48af90-823d-4250-9836-5dd843722eef service nova] Acquiring lock "054840f7-a1f2-4888-9a04-75c788dee325-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1304.717176] env[61987]: DEBUG oslo_concurrency.lockutils [req-a81410f9-ae3a-4248-a4a6-ebe6a756b96e req-4f48af90-823d-4250-9836-5dd843722eef service nova] Lock "054840f7-a1f2-4888-9a04-75c788dee325-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1304.717176] env[61987]: DEBUG oslo_concurrency.lockutils [req-a81410f9-ae3a-4248-a4a6-ebe6a756b96e req-4f48af90-823d-4250-9836-5dd843722eef service nova] Lock "054840f7-a1f2-4888-9a04-75c788dee325-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1304.719040] env[61987]: DEBUG nova.compute.manager [req-a81410f9-ae3a-4248-a4a6-ebe6a756b96e req-4f48af90-823d-4250-9836-5dd843722eef service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] No waiting events found dispatching network-vif-plugged-7906310e-5bc1-4ad3-85fb-062c271a2f85 {{(pid=61987) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1304.719263] env[61987]: WARNING nova.compute.manager [req-a81410f9-ae3a-4248-a4a6-ebe6a756b96e req-4f48af90-823d-4250-9836-5dd843722eef service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Received unexpected event network-vif-plugged-7906310e-5bc1-4ad3-85fb-062c271a2f85 for instance with vm_state shelved_offloaded and task_state spawning. [ 1304.807393] env[61987]: DEBUG oslo_concurrency.lockutils [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1304.807601] env[61987]: DEBUG oslo_concurrency.lockutils [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquired lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1304.807787] env[61987]: DEBUG nova.network.neutron [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Building network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1305.503463] env[61987]: DEBUG nova.network.neutron [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Updating instance_info_cache with network_info: [{"id": "7906310e-5bc1-4ad3-85fb-062c271a2f85", "address": "fa:16:3e:99:f5:8a", "network": {"id": "eac55241-aad5-4ddb-a01e-625c9298a8af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1411727225-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90faa728219a432fa241b635071729e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f996252-e329-42bd-a897-446dfe2b81cd", "external-id": "nsx-vlan-transportzone-535", "segmentation_id": 535, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7906310e-5b", "ovs_interfaceid": "7906310e-5bc1-4ad3-85fb-062c271a2f85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1306.007083] env[61987]: DEBUG oslo_concurrency.lockutils [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Releasing lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1306.033355] env[61987]: DEBUG nova.virt.hardware [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-30T23:05:45Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='62c42c78ee550fff599ac93a7884c3f8',container_format='bare',created_at=2024-09-30T23:20:02Z,direct_url=,disk_format='vmdk',id=e6a4516f-aab7-4816-a4c8-1b5f531bee5b,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-2046122420-shelved',owner='90faa728219a432fa241b635071729e5',properties=ImageMetaProps,protected=,size=31667200,status='active',tags=,updated_at=2024-09-30T23:20:15Z,virtual_size=,visibility=), allow threads: False {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1306.033620] env[61987]: DEBUG nova.virt.hardware [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Flavor limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1306.033789] env[61987]: DEBUG nova.virt.hardware [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Image limits 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1306.033985] env[61987]: DEBUG nova.virt.hardware [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Flavor pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1306.034161] env[61987]: DEBUG nova.virt.hardware [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Image pref 0:0:0 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1306.034318] env[61987]: DEBUG nova.virt.hardware [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61987) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1306.034530] env[61987]: DEBUG nova.virt.hardware [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1306.034696] env[61987]: DEBUG nova.virt.hardware [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1306.034868] env[61987]: DEBUG nova.virt.hardware [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Got 1 possible topologies {{(pid=61987) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1306.035048] env[61987]: DEBUG nova.virt.hardware [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1306.035236] env[61987]: DEBUG nova.virt.hardware [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61987) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1306.036106] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b100dee-f1b7-4ab5-bdf2-780af3531934 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.044072] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2887186f-18cd-4441-9f79-c3525d093d20 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.056743] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:f5:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1f996252-e329-42bd-a897-446dfe2b81cd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7906310e-5bc1-4ad3-85fb-062c271a2f85', 'vif_model': 'vmxnet3'}] {{(pid=61987) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1306.063959] env[61987]: DEBUG oslo.service.loopingcall [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1306.064203] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Creating VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1306.064406] env[61987]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cb1cb692-85a9-4c4f-a530-a8d43b7f4006 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.082208] env[61987]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1306.082208] env[61987]: value = "task-1062551" [ 1306.082208] env[61987]: _type = "Task" [ 1306.082208] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1306.089167] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062551, 'name': CreateVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1306.592556] env[61987]: DEBUG oslo_vmware.api [-] Task: {'id': task-1062551, 'name': CreateVM_Task, 'duration_secs': 0.288041} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1306.592784] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Created VM on the ESX host {{(pid=61987) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1306.599546] env[61987]: DEBUG oslo_concurrency.lockutils [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e6a4516f-aab7-4816-a4c8-1b5f531bee5b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1306.599716] env[61987]: DEBUG oslo_concurrency.lockutils [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e6a4516f-aab7-4816-a4c8-1b5f531bee5b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1306.600163] env[61987]: DEBUG oslo_concurrency.lockutils [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e6a4516f-aab7-4816-a4c8-1b5f531bee5b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1306.600409] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a4b064e-32b7-4651-8650-2da1e8bb5db6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1306.604467] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1306.604467] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]529b6774-4853-d7ef-5954-a3d8470070f8" [ 1306.604467] env[61987]: _type = "Task" [ 1306.604467] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1306.611414] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]529b6774-4853-d7ef-5954-a3d8470070f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1306.736960] env[61987]: DEBUG nova.compute.manager [req-9741f98d-6eb5-41ab-8a25-5a8322b65e91 req-c303522c-e57b-47e2-9ef4-5346998fe0cc service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Received event network-changed-7906310e-5bc1-4ad3-85fb-062c271a2f85 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1306.737167] env[61987]: DEBUG nova.compute.manager [req-9741f98d-6eb5-41ab-8a25-5a8322b65e91 req-c303522c-e57b-47e2-9ef4-5346998fe0cc service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Refreshing instance network info cache due to event network-changed-7906310e-5bc1-4ad3-85fb-062c271a2f85. {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11609}} [ 1306.737392] env[61987]: DEBUG oslo_concurrency.lockutils [req-9741f98d-6eb5-41ab-8a25-5a8322b65e91 req-c303522c-e57b-47e2-9ef4-5346998fe0cc service nova] Acquiring lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1306.737547] env[61987]: DEBUG oslo_concurrency.lockutils [req-9741f98d-6eb5-41ab-8a25-5a8322b65e91 req-c303522c-e57b-47e2-9ef4-5346998fe0cc service nova] Acquired lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1306.737720] env[61987]: DEBUG nova.network.neutron [req-9741f98d-6eb5-41ab-8a25-5a8322b65e91 req-c303522c-e57b-47e2-9ef4-5346998fe0cc service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Refreshing network info cache for port 7906310e-5bc1-4ad3-85fb-062c271a2f85 {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1307.114663] env[61987]: DEBUG oslo_concurrency.lockutils [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e6a4516f-aab7-4816-a4c8-1b5f531bee5b" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1307.114978] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Processing image e6a4516f-aab7-4816-a4c8-1b5f531bee5b {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1307.115141] env[61987]: DEBUG oslo_concurrency.lockutils [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e6a4516f-aab7-4816-a4c8-1b5f531bee5b/e6a4516f-aab7-4816-a4c8-1b5f531bee5b.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1307.115300] env[61987]: DEBUG oslo_concurrency.lockutils [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e6a4516f-aab7-4816-a4c8-1b5f531bee5b/e6a4516f-aab7-4816-a4c8-1b5f531bee5b.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1307.115481] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1307.115725] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3b56a548-929d-4076-881d-877d4e2ccf4c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.123389] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1307.123567] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61987) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1307.124225] env[61987]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6fdef64c-a553-4963-abfa-4aa84f11cfb6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.128814] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1307.128814] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]52cb92c6-9e00-9ac1-99b6-801deaed4ef7" [ 1307.128814] env[61987]: _type = "Task" [ 1307.128814] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.135662] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': session[5200fa56-000c-905f-584b-301f4385e440]52cb92c6-9e00-9ac1-99b6-801deaed4ef7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.427285] env[61987]: DEBUG nova.network.neutron [req-9741f98d-6eb5-41ab-8a25-5a8322b65e91 req-c303522c-e57b-47e2-9ef4-5346998fe0cc service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Updated VIF entry in instance network info cache for port 7906310e-5bc1-4ad3-85fb-062c271a2f85. {{(pid=61987) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1307.427671] env[61987]: DEBUG nova.network.neutron [req-9741f98d-6eb5-41ab-8a25-5a8322b65e91 req-c303522c-e57b-47e2-9ef4-5346998fe0cc service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Updating instance_info_cache with network_info: [{"id": "7906310e-5bc1-4ad3-85fb-062c271a2f85", "address": "fa:16:3e:99:f5:8a", "network": {"id": "eac55241-aad5-4ddb-a01e-625c9298a8af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1411727225-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90faa728219a432fa241b635071729e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f996252-e329-42bd-a897-446dfe2b81cd", "external-id": "nsx-vlan-transportzone-535", "segmentation_id": 535, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7906310e-5b", "ovs_interfaceid": "7906310e-5bc1-4ad3-85fb-062c271a2f85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1307.639045] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Preparing fetch location {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1307.639330] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Fetch image to [datastore2] OSTACK_IMG_56bbec89-43c3-4dd2-b411-ef6623a850cc/OSTACK_IMG_56bbec89-43c3-4dd2-b411-ef6623a850cc.vmdk {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1307.639523] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Downloading stream optimized image e6a4516f-aab7-4816-a4c8-1b5f531bee5b to [datastore2] OSTACK_IMG_56bbec89-43c3-4dd2-b411-ef6623a850cc/OSTACK_IMG_56bbec89-43c3-4dd2-b411-ef6623a850cc.vmdk on the data store datastore2 as vApp {{(pid=61987) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1307.639701] env[61987]: DEBUG nova.virt.vmwareapi.images [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Downloading image file data e6a4516f-aab7-4816-a4c8-1b5f531bee5b to the ESX as VM named 'OSTACK_IMG_56bbec89-43c3-4dd2-b411-ef6623a850cc' {{(pid=61987) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1307.704697] env[61987]: DEBUG oslo_vmware.rw_handles [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1307.704697] env[61987]: value = "resgroup-9" [ 1307.704697] env[61987]: _type = "ResourcePool" [ 1307.704697] env[61987]: }. {{(pid=61987) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1307.704978] env[61987]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-1281cfe7-b14d-4d21-9509-ef24088e348b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.724744] env[61987]: DEBUG oslo_vmware.rw_handles [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lease: (returnval){ [ 1307.724744] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5287805b-9e0a-ceda-4fb4-5b2433bcc2b9" [ 1307.724744] env[61987]: _type = "HttpNfcLease" [ 1307.724744] env[61987]: } obtained for vApp import into resource pool (val){ [ 1307.724744] env[61987]: value = "resgroup-9" [ 1307.724744] env[61987]: _type = "ResourcePool" [ 1307.724744] env[61987]: }. {{(pid=61987) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1307.725037] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the lease: (returnval){ [ 1307.725037] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5287805b-9e0a-ceda-4fb4-5b2433bcc2b9" [ 1307.725037] env[61987]: _type = "HttpNfcLease" [ 1307.725037] env[61987]: } to be ready. {{(pid=61987) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1307.731392] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1307.731392] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5287805b-9e0a-ceda-4fb4-5b2433bcc2b9" [ 1307.731392] env[61987]: _type = "HttpNfcLease" [ 1307.731392] env[61987]: } is initializing. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1307.930827] env[61987]: DEBUG oslo_concurrency.lockutils [req-9741f98d-6eb5-41ab-8a25-5a8322b65e91 req-c303522c-e57b-47e2-9ef4-5346998fe0cc service nova] Releasing lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1308.233464] env[61987]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1308.233464] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5287805b-9e0a-ceda-4fb4-5b2433bcc2b9" [ 1308.233464] env[61987]: _type = "HttpNfcLease" [ 1308.233464] env[61987]: } is ready. {{(pid=61987) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1308.233908] env[61987]: DEBUG oslo_vmware.rw_handles [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1308.233908] env[61987]: value = "session[5200fa56-000c-905f-584b-301f4385e440]5287805b-9e0a-ceda-4fb4-5b2433bcc2b9" [ 1308.233908] env[61987]: _type = "HttpNfcLease" [ 1308.233908] env[61987]: }. {{(pid=61987) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1308.234456] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ba966f-9543-4b8e-8b8a-0e0978b6e8e5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.241528] env[61987]: DEBUG oslo_vmware.rw_handles [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e55cc9-919f-dbe6-082f-cafac80bc992/disk-0.vmdk from lease info. {{(pid=61987) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1308.241701] env[61987]: DEBUG oslo_vmware.rw_handles [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Creating HTTP connection to write to file with size = 31667200 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e55cc9-919f-dbe6-082f-cafac80bc992/disk-0.vmdk. {{(pid=61987) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1308.304568] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2720ea36-b5e8-4260-8169-b68cc157a32a {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.455267] env[61987]: DEBUG oslo_vmware.rw_handles [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Completed reading data from the image iterator. {{(pid=61987) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1309.455676] env[61987]: DEBUG oslo_vmware.rw_handles [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e55cc9-919f-dbe6-082f-cafac80bc992/disk-0.vmdk. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1309.456560] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e8efcc-4d04-40ae-9667-60078e753e1d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.463596] env[61987]: DEBUG oslo_vmware.rw_handles [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e55cc9-919f-dbe6-082f-cafac80bc992/disk-0.vmdk is in state: ready. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1309.463808] env[61987]: DEBUG oslo_vmware.rw_handles [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e55cc9-919f-dbe6-082f-cafac80bc992/disk-0.vmdk. {{(pid=61987) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1309.464061] env[61987]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-f0d8935e-7b8a-467d-b682-2668b8ffac33 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.632646] env[61987]: DEBUG oslo_vmware.rw_handles [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e55cc9-919f-dbe6-082f-cafac80bc992/disk-0.vmdk. {{(pid=61987) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1309.632876] env[61987]: INFO nova.virt.vmwareapi.images [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Downloaded image file data e6a4516f-aab7-4816-a4c8-1b5f531bee5b [ 1309.633878] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab70be3d-54fc-419a-b929-d07ca957f203 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.650831] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-34586155-3001-4dde-9798-8da8e432640d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.693151] env[61987]: INFO nova.virt.vmwareapi.images [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] The imported VM was unregistered [ 1309.695363] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Caching image {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1309.695604] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Creating directory with path [datastore2] devstack-image-cache_base/e6a4516f-aab7-4816-a4c8-1b5f531bee5b {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1309.695872] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5a0cac91-65ba-4ef6-a46d-c1b33adce560 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.705381] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Created directory with path [datastore2] devstack-image-cache_base/e6a4516f-aab7-4816-a4c8-1b5f531bee5b {{(pid=61987) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1309.705565] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_56bbec89-43c3-4dd2-b411-ef6623a850cc/OSTACK_IMG_56bbec89-43c3-4dd2-b411-ef6623a850cc.vmdk to [datastore2] devstack-image-cache_base/e6a4516f-aab7-4816-a4c8-1b5f531bee5b/e6a4516f-aab7-4816-a4c8-1b5f531bee5b.vmdk. {{(pid=61987) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1309.705802] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-4f4a1bcc-0e19-49d8-80cd-ab532870259d {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.712589] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1309.712589] env[61987]: value = "task-1062554" [ 1309.712589] env[61987]: _type = "Task" [ 1309.712589] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1309.719876] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062554, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1310.224472] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062554, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1310.724980] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062554, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1311.227921] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062554, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1311.726323] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062554, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1312.224891] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062554, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.283342} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1312.225191] env[61987]: INFO nova.virt.vmwareapi.ds_util [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_56bbec89-43c3-4dd2-b411-ef6623a850cc/OSTACK_IMG_56bbec89-43c3-4dd2-b411-ef6623a850cc.vmdk to [datastore2] devstack-image-cache_base/e6a4516f-aab7-4816-a4c8-1b5f531bee5b/e6a4516f-aab7-4816-a4c8-1b5f531bee5b.vmdk. [ 1312.225388] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Cleaning up location [datastore2] OSTACK_IMG_56bbec89-43c3-4dd2-b411-ef6623a850cc {{(pid=61987) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1312.225557] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_56bbec89-43c3-4dd2-b411-ef6623a850cc {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1312.225809] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fff787f0-01ea-458d-b6c3-f225371c564f {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.232443] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1312.232443] env[61987]: value = "task-1062555" [ 1312.232443] env[61987]: _type = "Task" [ 1312.232443] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1312.239916] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062555, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1312.743108] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062555, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.037443} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1312.743511] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1312.743607] env[61987]: DEBUG oslo_concurrency.lockutils [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e6a4516f-aab7-4816-a4c8-1b5f531bee5b/e6a4516f-aab7-4816-a4c8-1b5f531bee5b.vmdk" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1312.743781] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e6a4516f-aab7-4816-a4c8-1b5f531bee5b/e6a4516f-aab7-4816-a4c8-1b5f531bee5b.vmdk to [datastore2] 054840f7-a1f2-4888-9a04-75c788dee325/054840f7-a1f2-4888-9a04-75c788dee325.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1312.744032] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b68da510-530f-4504-92bf-f6cd7c69e820 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.750238] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1312.750238] env[61987]: value = "task-1062556" [ 1312.750238] env[61987]: _type = "Task" [ 1312.750238] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1312.757139] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062556, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1313.262215] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062556, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1313.764258] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062556, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1314.264297] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062556, 'name': CopyVirtualDisk_Task} progress is 71%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1314.764980] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062556, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1315.263647] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062556, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.08183} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1315.263917] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e6a4516f-aab7-4816-a4c8-1b5f531bee5b/e6a4516f-aab7-4816-a4c8-1b5f531bee5b.vmdk to [datastore2] 054840f7-a1f2-4888-9a04-75c788dee325/054840f7-a1f2-4888-9a04-75c788dee325.vmdk {{(pid=61987) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1315.264696] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59760e73-279f-4e13-a1cb-d933827ecda6 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.286086] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Reconfiguring VM instance instance-00000077 to attach disk [datastore2] 054840f7-a1f2-4888-9a04-75c788dee325/054840f7-a1f2-4888-9a04-75c788dee325.vmdk or device None with type streamOptimized {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1315.286313] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dfba9737-49bf-4a19-85c0-03bbfa7418c0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.304607] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1315.304607] env[61987]: value = "task-1062557" [ 1315.304607] env[61987]: _type = "Task" [ 1315.304607] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1315.312407] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062557, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1315.815104] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062557, 'name': ReconfigVM_Task, 'duration_secs': 0.288535} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1315.815425] env[61987]: DEBUG nova.virt.vmwareapi.volumeops [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Reconfigured VM instance instance-00000077 to attach disk [datastore2] 054840f7-a1f2-4888-9a04-75c788dee325/054840f7-a1f2-4888-9a04-75c788dee325.vmdk or device None with type streamOptimized {{(pid=61987) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1315.815970] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f1db1322-0879-4d17-98c1-db564ba88ef7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.821899] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1315.821899] env[61987]: value = "task-1062558" [ 1315.821899] env[61987]: _type = "Task" [ 1315.821899] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1315.828864] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062558, 'name': Rename_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1316.331494] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062558, 'name': Rename_Task, 'duration_secs': 0.134336} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1316.331776] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Powering on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1316.332033] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e203c3a-c5b2-4ce4-a74a-218f06d3487b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.337977] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1316.337977] env[61987]: value = "task-1062559" [ 1316.337977] env[61987]: _type = "Task" [ 1316.337977] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1316.344805] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062559, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1316.847298] env[61987]: DEBUG oslo_vmware.api [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062559, 'name': PowerOnVM_Task, 'duration_secs': 0.415927} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1316.847610] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Powered on the VM {{(pid=61987) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1316.939242] env[61987]: DEBUG nova.compute.manager [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Checking state {{(pid=61987) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1316.940194] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49e83e7e-3ef7-41d8-b012-c35e5986d2e1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.458712] env[61987]: DEBUG oslo_concurrency.lockutils [None req-620536df-381c-47a7-aa9a-77ed4950990f tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "054840f7-a1f2-4888-9a04-75c788dee325" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.326s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1319.650791] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1319.651228] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1319.651273] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61987) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11020}} [ 1319.961078] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1319.964647] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1320.965411] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1320.965934] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Starting heal instance info cache {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10401}} [ 1320.965934] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Rebuilding the list of instances to heal {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10405}} [ 1321.496251] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1321.496433] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquired lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1321.496617] env[61987]: DEBUG nova.network.neutron [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Forcefully refreshing network info cache for instance {{(pid=61987) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1321.496777] env[61987]: DEBUG nova.objects.instance [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lazy-loading 'info_cache' on Instance uuid 054840f7-a1f2-4888-9a04-75c788dee325 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1323.213691] env[61987]: DEBUG nova.network.neutron [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Updating instance_info_cache with network_info: [{"id": "7906310e-5bc1-4ad3-85fb-062c271a2f85", "address": "fa:16:3e:99:f5:8a", "network": {"id": "eac55241-aad5-4ddb-a01e-625c9298a8af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1411727225-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "90faa728219a432fa241b635071729e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1f996252-e329-42bd-a897-446dfe2b81cd", "external-id": "nsx-vlan-transportzone-535", "segmentation_id": 535, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7906310e-5b", "ovs_interfaceid": "7906310e-5bc1-4ad3-85fb-062c271a2f85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1323.716015] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Releasing lock "refresh_cache-054840f7-a1f2-4888-9a04-75c788dee325" {{(pid=61987) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1323.716326] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Updated the network info_cache for instance {{(pid=61987) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10472}} [ 1323.716567] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1323.716740] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1323.716902] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager.update_available_resource {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1324.220125] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1324.220572] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1324.220572] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1324.220703] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61987) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1324.221658] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd949e4b-9cae-498c-ab23-089f8949758c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.230942] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1926f95c-532e-4afa-9f4a-ea4fdffaaee1 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.244774] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9915b3b-def2-4431-aef9-8cad9dd04df5 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.250928] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-772e0f6a-2acc-4050-b0ca-ea3682717352 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.279329] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181588MB free_disk=179GB free_vcpus=48 pci_devices=None {{(pid=61987) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1324.279494] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1324.279714] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1325.302854] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Instance 054840f7-a1f2-4888-9a04-75c788dee325 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61987) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1325.303134] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1325.303221] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61987) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1325.329495] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f53bf763-4039-4a6d-a9fe-023f62b509e0 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.336830] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68a25a9c-98ab-4c9c-8c48-e26d44cd03d9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.366555] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a02d4032-2bb1-4f8a-8c19-2c84edadd2de {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.373011] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33ce8047-7f1e-4923-879c-27bc9df74eff {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.385422] env[61987]: DEBUG nova.compute.provider_tree [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1325.904759] env[61987]: ERROR nova.scheduler.client.report [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [req-1a2f6389-7a2c-45ab-a3c7-3524abd671d6] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 8db744e5-1e14-4e98-9818-695ea4b7dfd0. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-1a2f6389-7a2c-45ab-a3c7-3524abd671d6"}]} [ 1325.920562] env[61987]: DEBUG nova.scheduler.client.report [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Refreshing inventories for resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 {{(pid=61987) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1325.934092] env[61987]: DEBUG nova.scheduler.client.report [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Updating ProviderTree inventory for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1325.934288] env[61987]: DEBUG nova.compute.provider_tree [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1325.944878] env[61987]: DEBUG nova.scheduler.client.report [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Refreshing aggregate associations for resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0, aggregates: None {{(pid=61987) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1325.964011] env[61987]: DEBUG nova.scheduler.client.report [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Refreshing trait associations for resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61987) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1325.988551] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0463fcf4-c85a-4483-b5fb-30ccb2c498b3 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.995774] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d58f5a5-cdff-4121-9d7b-005e208d9279 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.024031] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab36ca0-2304-493a-b24a-af55d73ba264 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.030838] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b096c87a-743c-4afa-a735-e85435f59d78 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.044124] env[61987]: DEBUG nova.compute.provider_tree [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1326.573559] env[61987]: DEBUG nova.scheduler.client.report [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Updated inventory for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with generation 158 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1326.573894] env[61987]: DEBUG nova.compute.provider_tree [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Updating resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 generation from 158 to 159 during operation: update_inventory {{(pid=61987) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1326.573955] env[61987]: DEBUG nova.compute.provider_tree [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 179, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1327.078798] env[61987]: DEBUG nova.compute.resource_tracker [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61987) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1327.079023] env[61987]: DEBUG oslo_concurrency.lockutils [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.799s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1327.327460] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1330.961085] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1353.703884] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "054840f7-a1f2-4888-9a04-75c788dee325" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1353.704297] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "054840f7-a1f2-4888-9a04-75c788dee325" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1353.704395] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "054840f7-a1f2-4888-9a04-75c788dee325-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1353.704553] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "054840f7-a1f2-4888-9a04-75c788dee325-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1353.704732] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "054840f7-a1f2-4888-9a04-75c788dee325-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1353.708389] env[61987]: INFO nova.compute.manager [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Terminating instance [ 1354.212129] env[61987]: DEBUG nova.compute.manager [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Start destroying the instance on the hypervisor. {{(pid=61987) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1354.212406] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Destroying instance {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1354.213454] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72ec26ec-f2b1-430e-8b69-a709a8a3ee2c {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.221602] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Powering off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1354.221829] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-744797a9-32a3-48c2-9683-bb767c196276 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.228588] env[61987]: DEBUG oslo_vmware.api [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1354.228588] env[61987]: value = "task-1062560" [ 1354.228588] env[61987]: _type = "Task" [ 1354.228588] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1354.235988] env[61987]: DEBUG oslo_vmware.api [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062560, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1354.738614] env[61987]: DEBUG oslo_vmware.api [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062560, 'name': PowerOffVM_Task, 'duration_secs': 0.168364} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1354.738979] env[61987]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Powered off the VM {{(pid=61987) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1354.739050] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Unregistering the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1354.739245] env[61987]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fdd9dfba-6dab-4f95-9d6d-f5b426ac48df {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.801383] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Unregistered the VM {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1354.801630] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Deleting contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1354.801819] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Deleting the datastore file [datastore2] 054840f7-a1f2-4888-9a04-75c788dee325 {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1354.802092] env[61987]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e910bab9-43b6-40d8-9899-ceae9b2250ae {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1354.807915] env[61987]: DEBUG oslo_vmware.api [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for the task: (returnval){ [ 1354.807915] env[61987]: value = "task-1062562" [ 1354.807915] env[61987]: _type = "Task" [ 1354.807915] env[61987]: } to complete. {{(pid=61987) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1354.814953] env[61987]: DEBUG oslo_vmware.api [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062562, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1355.322150] env[61987]: DEBUG oslo_vmware.api [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Task: {'id': task-1062562, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.114315} completed successfully. {{(pid=61987) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1355.322567] env[61987]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Deleted the datastore file {{(pid=61987) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1355.322867] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Deleted contents of the VM from datastore datastore2 {{(pid=61987) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1355.323173] env[61987]: DEBUG nova.virt.vmwareapi.vmops [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Instance destroyed {{(pid=61987) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1355.323483] env[61987]: INFO nova.compute.manager [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1355.323865] env[61987]: DEBUG oslo.service.loopingcall [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61987) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1355.324172] env[61987]: DEBUG nova.compute.manager [-] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Deallocating network for instance {{(pid=61987) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1355.324321] env[61987]: DEBUG nova.network.neutron [-] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] deallocate_for_instance() {{(pid=61987) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1355.751956] env[61987]: DEBUG nova.compute.manager [req-94d7f7ea-7035-499a-a437-ada8302e03e4 req-036d340c-a146-4ef7-ba6f-5fa10f06e73a service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Received event network-vif-deleted-7906310e-5bc1-4ad3-85fb-062c271a2f85 {{(pid=61987) external_instance_event /opt/stack/nova/nova/compute/manager.py:11604}} [ 1355.752336] env[61987]: INFO nova.compute.manager [req-94d7f7ea-7035-499a-a437-ada8302e03e4 req-036d340c-a146-4ef7-ba6f-5fa10f06e73a service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Neutron deleted interface 7906310e-5bc1-4ad3-85fb-062c271a2f85; detaching it from the instance and deleting it from the info cache [ 1355.752336] env[61987]: DEBUG nova.network.neutron [req-94d7f7ea-7035-499a-a437-ada8302e03e4 req-036d340c-a146-4ef7-ba6f-5fa10f06e73a service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1356.226051] env[61987]: DEBUG nova.network.neutron [-] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Updating instance_info_cache with network_info: [] {{(pid=61987) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1356.254675] env[61987]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b8f039e3-8e2a-40d7-8f14-34cb804cfd28 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.263969] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74091a3a-a82e-4155-a6dd-cdef566c89fd {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.287654] env[61987]: DEBUG nova.compute.manager [req-94d7f7ea-7035-499a-a437-ada8302e03e4 req-036d340c-a146-4ef7-ba6f-5fa10f06e73a service nova] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Detach interface failed, port_id=7906310e-5bc1-4ad3-85fb-062c271a2f85, reason: Instance 054840f7-a1f2-4888-9a04-75c788dee325 could not be found. {{(pid=61987) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11438}} [ 1356.729073] env[61987]: INFO nova.compute.manager [-] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Took 1.40 seconds to deallocate network for instance. [ 1357.235496] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1357.235875] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1357.236065] env[61987]: DEBUG nova.objects.instance [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lazy-loading 'resources' on Instance uuid 054840f7-a1f2-4888-9a04-75c788dee325 {{(pid=61987) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1357.767930] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67b22f7-1d83-4d28-bd22-9ec102e348f7 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.775668] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff858fa-fd00-453a-a829-33513295aaa9 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.805721] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e43b15a7-6a04-4264-b044-0ec9612ae351 {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.812958] env[61987]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d0d587-e8f1-4d4d-87f4-959d4ce9b35b {{(pid=61987) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1357.825838] env[61987]: DEBUG nova.compute.provider_tree [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1358.357120] env[61987]: DEBUG nova.scheduler.client.report [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Updated inventory for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with generation 159 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1358.357516] env[61987]: DEBUG nova.compute.provider_tree [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Updating resource provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 generation from 159 to 160 during operation: update_inventory {{(pid=61987) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1358.357591] env[61987]: DEBUG nova.compute.provider_tree [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Updating inventory in ProviderTree for provider 8db744e5-1e14-4e98-9818-695ea4b7dfd0 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 180, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61987) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1358.862119] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.626s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1358.886677] env[61987]: INFO nova.scheduler.client.report [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Deleted allocations for instance 054840f7-a1f2-4888-9a04-75c788dee325 [ 1359.395225] env[61987]: DEBUG oslo_concurrency.lockutils [None req-d5cd1f4c-9eec-4b86-b1cc-9c1e1230ff51 tempest-AttachVolumeShelveTestJSON-1998204518 tempest-AttachVolumeShelveTestJSON-1998204518-project-member] Lock "054840f7-a1f2-4888-9a04-75c788dee325" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.691s {{(pid=61987) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1378.964805] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1378.965267] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1378.965267] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61987) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11020}} [ 1379.961916] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1379.965571] env[61987]: DEBUG oslo_service.periodic_task [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61987) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1379.965955] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] Cleaning up deleted instances {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11701}} [ 1380.472828] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] There are 18 instances to clean {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11710}} [ 1380.473076] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 03766f52-c102-40ff-b004-7748231d0a03] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1380.976402] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 9c5957a4-e7a1-4ac2-85d8-f7bc394eea5f] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1381.480374] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 054840f7-a1f2-4888-9a04-75c788dee325] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1381.984618] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: b64111ab-4cd4-4584-b64e-e61b129a721d] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1382.488230] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 0704efaf-08de-4888-a092-476bcb789eea] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1382.991375] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 430c06b5-8e87-4fe3-bdc7-3fd7b936dcd3] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1383.495018] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 2ba77d32-d777-448e-8fa6-04920950f0f6] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1383.998635] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 5912b687-2150-4133-a3a1-44ce38b3c3cb] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1384.502400] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: f030044b-bc6b-478c-8f63-d519aa76203c] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1385.005384] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 76707afe-29ae-4257-a237-ba5a47b2511b] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1385.509816] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 7002962c-76ed-4a8d-be2a-b2b118847fea] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1386.013458] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: bf1123bf-80e6-4be3-804f-e13b906ee44d] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}} [ 1386.517303] env[61987]: DEBUG nova.compute.manager [None req-463db6b5-852b-47ad-a7c9-318a36539757 None None] [instance: 0cba5e51-7e0a-409b-a7b2-612ab8cf5689] Instance has had 0 of 5 cleanup attempts {{(pid=61987) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11714}}